1. 16 Oct, 2020 4 commits
    • Stas Bekman's avatar
      fix/hide warnings (#7837) · d8ca57d2
      Stas Bekman authored
      s
      d8ca57d2
    • Sam Shleifer's avatar
      [cleanup] assign todos, faster bart-cnn test (#7835) · 96e47d92
      Sam Shleifer authored
      * 2 beam output
      
      * unassign/remove TODOs
      
      * remove one more
      96e47d92
    • rmroczkowski's avatar
      Herbert polish model (#7798) · 7b13bd01
      rmroczkowski authored
      
      
      * HerBERT transformer model for Polish language understanding.
      
      * HerbertTokenizerFast generated with HerbertConverter
      
      * Herbert base and large model cards
      
      * Herbert model cards with tags
      
      * Herbert tensorflow models
      
      * Herbert model tests based on Bert test suit
      
      * src/transformers/tokenization_herbert.py edited online with Bitbucket
      
      * src/transformers/tokenization_herbert.py edited online with Bitbucket
      
      * docs/source/model_doc/herbert.rst edited online with Bitbucket
      
      * Herbert tokenizer tests and bug fixes
      
      * src/transformers/configuration_herbert.py edited online with Bitbucket
      
      * Copyrights and tests for TFHerbertModel
      
      * model_cards/allegro/herbert-base-cased/README.md edited online with Bitbucket
      
      * model_cards/allegro/herbert-large-cased/README.md edited online with Bitbucket
      
      * Bug fixes after testing
      
      * Reformat modified_only_fixup
      
      * Proper order of configuration
      
      * Herbert proper documentation formatting
      
      * Formatting with make modified_only_fixup
      
      * Dummies fixed
      
      * Adding missing models to documentation
      
      * Removing HerBERT model as it is a simple extension of BERT
      
      * Update model_cards/allegro/herbert-base-cased/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      
      * Update model_cards/allegro/herbert-large-cased/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      
      * HerbertTokenizer deprecated configuration removed
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      7b13bd01
    • Lysandre Debut's avatar
      Fix DeBERTa integration tests (#7729) · 52c9e842
      Lysandre Debut authored
      52c9e842
  2. 15 Oct, 2020 1 commit
  3. 14 Oct, 2020 2 commits
  4. 13 Oct, 2020 4 commits
  5. 10 Oct, 2020 1 commit
  6. 09 Oct, 2020 2 commits
  7. 08 Oct, 2020 2 commits
    • Thomas Wolf's avatar
      Adding Fast tokenizers for SentencePiece based tokenizers - Breaking: remove... · 9aeacb58
      Thomas Wolf authored
      
      Adding Fast tokenizers for SentencePiece based tokenizers - Breaking: remove Transfo-XL fast tokenizer (#7141)
      
      * [WIP] SP tokenizers
      
      * fixing tests for T5
      
      * WIP tokenizers
      
      * serialization
      
      * update T5
      
      * WIP T5 tokenization
      
      * slow to fast conversion script
      
      * Refactoring to move tokenzier implementations inside transformers
      
      * Adding gpt - refactoring - quality
      
      * WIP adding several tokenizers to the fast world
      
      * WIP Roberta - moving implementations
      
      * update to dev4 switch file loading to in-memory loading
      
      * Updating and fixing
      
      * advancing on the tokenizers - updating do_lower_case
      
      * style and quality
      
      * moving forward with tokenizers conversion and tests
      
      * MBart, T5
      
      * dumping the fast version of transformer XL
      
      * Adding to autotokenizers + style/quality
      
      * update init and space_between_special_tokens
      
      * style and quality
      
      * bump up tokenizers version
      
      * add protobuf
      
      * fix pickle Bert JP with Mecab
      
      * fix newly added tokenizers
      
      * style and quality
      
      * fix bert japanese
      
      * fix funnel
      
      * limite tokenizer warning to one occurence
      
      * clean up file
      
      * fix new tokenizers
      
      * fast tokenizers deep tests
      
      * WIP adding all the special fast tests on the new fast tokenizers
      
      * quick fix
      
      * adding more fast tokenizers in the fast tests
      
      * all tokenizers in fast version tested
      
      * Adding BertGenerationFast
      
      * bump up setup.py for CI
      
      * remove BertGenerationFast (too early)
      
      * bump up tokenizers version
      
      * Clean old docstrings
      
      * Typo
      
      * Update following Lysandre comments
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      9aeacb58
    • Sam Shleifer's avatar
      e3e65173
  8. 07 Oct, 2020 2 commits
  9. 06 Oct, 2020 1 commit
  10. 05 Oct, 2020 3 commits
    • Julien Plu's avatar
      Custom TF weights loading (#7422) · 9cf7b23b
      Julien Plu authored
      
      
      * First try
      
      * Fix TF utils
      
      * Handle authorized unexpected keys when loading weights
      
      * Add several more authorized unexpected keys
      
      * Apply style
      
      * Fix test
      
      * Address Patrick's comments.
      
      * Update src/transformers/modeling_tf_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/modeling_tf_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply style
      
      * Make return_dict the default behavior and display a warning message
      
      * Revert
      
      * Replace wrong keyword
      
      * Revert code
      
      * Add forgot key
      
      * Fix bug in loading PT models from a TF one.
      
      * Fix sort
      
      * Add a test for custom load weights in BERT
      
      * Apply style
      
      * Remove unused import
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      9cf7b23b
    • Sylvain Gugger's avatar
      Expand test to locate flakiness (#7580) · d3adb985
      Sylvain Gugger authored
      d3adb985
    • Forrest Iandola's avatar
      SqueezeBERT architecture (#7083) · 02ef825b
      Forrest Iandola authored
      * configuration_squeezebert.py
      
      thin wrapper around bert tokenizer
      
      fix typos
      
      wip sb model code
      
      wip modeling_squeezebert.py. Next step is to get the multi-layer-output interface working
      
      set up squeezebert to use BertModelOutput when returning results.
      
      squeezebert documentation
      
      formatting
      
      allow head mask that is an array of [None, ..., None]
      
      docs
      
      docs cont'd
      
      path to vocab
      
      docs and pointers to cloud files (WIP)
      
      line length and indentation
      
      squeezebert model cards
      
      formatting of model cards
      
      untrack modeling_squeezebert_scratchpad.py
      
      update aws paths to vocab and config files
      
      get rid of stub of NSP code, and advise users to pretrain with mlm only
      
      fix rebase issues
      
      redo rebase of modeling_auto.py
      
      fix issues with code formatting
      
      more code format auto-fixes
      
      move squeezebert before bert in tokenization_auto.py and modeling_auto.py because squeezebert inherits from bert
      
      tests for squeezebert modeling and tokenization
      
      fix typo
      
      move squeezebert before bert in modeling_auto.py to fix inheritance problem
      
      disable test_head_masking, since squeezebert doesn't yet implement head masking
      
      fix issues exposed by the test_modeling_squeezebert.py
      
      fix an issue exposed by test_tokenization_squeezebert.py
      
      fix issue exposed by test_modeling_squeezebert.py
      
      auto generated code style improvement
      
      issue that we inherited from modeling_xxx.py: SqueezeBertForMaskedLM.forward() calls self.cls(), but there is no self.cls, and I think the goal was actually to call self.lm_head()
      
      update copyright
      
      resolve failing 'test_hidden_states_output' and remove unused encoder_hidden_states and encoder_attention_mask
      
      docs
      
      add integration test. rename squeezebert-mnli --> squeezebert/squeezebert-mnli
      
      autogenerated formatting tweaks
      
      integrate feedback from patrickvonplaten and sgugger to programming style and documentation strings
      
      * tiny change to order of imports
      02ef825b
  11. 01 Oct, 2020 3 commits
  12. 30 Sep, 2020 2 commits
  13. 29 Sep, 2020 3 commits
  14. 28 Sep, 2020 2 commits
  15. 25 Sep, 2020 5 commits
  16. 24 Sep, 2020 2 commits
  17. 23 Sep, 2020 1 commit