1. 16 Oct, 2020 2 commits
    • Patrick von Platen's avatar
      Fix typo in sequence model card · dc552b9b
      Patrick von Platen authored
      dc552b9b
    • rmroczkowski's avatar
      Herbert polish model (#7798) · 7b13bd01
      rmroczkowski authored
      
      
      * HerBERT transformer model for Polish language understanding.
      
      * HerbertTokenizerFast generated with HerbertConverter
      
      * Herbert base and large model cards
      
      * Herbert model cards with tags
      
      * Herbert tensorflow models
      
      * Herbert model tests based on Bert test suit
      
      * src/transformers/tokenization_herbert.py edited online with Bitbucket
      
      * src/transformers/tokenization_herbert.py edited online with Bitbucket
      
      * docs/source/model_doc/herbert.rst edited online with Bitbucket
      
      * Herbert tokenizer tests and bug fixes
      
      * src/transformers/configuration_herbert.py edited online with Bitbucket
      
      * Copyrights and tests for TFHerbertModel
      
      * model_cards/allegro/herbert-base-cased/README.md edited online with Bitbucket
      
      * model_cards/allegro/herbert-large-cased/README.md edited online with Bitbucket
      
      * Bug fixes after testing
      
      * Reformat modified_only_fixup
      
      * Proper order of configuration
      
      * Herbert proper documentation formatting
      
      * Formatting with make modified_only_fixup
      
      * Dummies fixed
      
      * Adding missing models to documentation
      
      * Removing HerBERT model as it is a simple extension of BERT
      
      * Update model_cards/allegro/herbert-base-cased/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      
      * Update model_cards/allegro/herbert-large-cased/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      
      * HerbertTokenizer deprecated configuration removed
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      7b13bd01
  2. 15 Oct, 2020 4 commits
  3. 14 Oct, 2020 6 commits
  4. 12 Oct, 2020 1 commit
  5. 10 Oct, 2020 1 commit
  6. 09 Oct, 2020 1 commit
  7. 07 Oct, 2020 9 commits
  8. 06 Oct, 2020 4 commits
  9. 05 Oct, 2020 3 commits
    • Joshua H's avatar
      Update Code example according to deprecation of AutoModeWithLMHead (#7555) · 1a00f46c
      Joshua H authored
      'The class `AutoModelWithLMHead` is deprecated and will be removed in a future version. Please use `AutoModelForCausalLM` for causal language models, `AutoModelForMaskedLM` for masked language models and `AutoModelForSeq2SeqLM` for encoder-decoder models.'
      I dont know how to change the 'How to use this model directly from the 馃/transformers library:' part since it is not part of the model-paper
      1a00f46c
    • Nathan Cooper's avatar
      [Model card] Java Code Summarizer model (#7568) · 071970fe
      Nathan Cooper authored
      
      
      * Create README.md
      
      * Update model_cards/ncoop57/bart-base-code-summarizer-java-v0/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      071970fe
    • Forrest Iandola's avatar
      SqueezeBERT architecture (#7083) · 02ef825b
      Forrest Iandola authored
      * configuration_squeezebert.py
      
      thin wrapper around bert tokenizer
      
      fix typos
      
      wip sb model code
      
      wip modeling_squeezebert.py. Next step is to get the multi-layer-output interface working
      
      set up squeezebert to use BertModelOutput when returning results.
      
      squeezebert documentation
      
      formatting
      
      allow head mask that is an array of [None, ..., None]
      
      docs
      
      docs cont'd
      
      path to vocab
      
      docs and pointers to cloud files (WIP)
      
      line length and indentation
      
      squeezebert model cards
      
      formatting of model cards
      
      untrack modeling_squeezebert_scratchpad.py
      
      update aws paths to vocab and config files
      
      get rid of stub of NSP code, and advise users to pretrain with mlm only
      
      fix rebase issues
      
      redo rebase of modeling_auto.py
      
      fix issues with code formatting
      
      more code format auto-fixes
      
      move squeezebert before bert in tokenization_auto.py and modeling_auto.py because squeezebert inherits from bert
      
      tests for squeezebert modeling and tokenization
      
      fix typo
      
      move squeezebert before bert in modeling_auto.py to fix inheritance problem
      
      disable test_head_masking, since squeezebert doesn't yet implement head masking
      
      fix issues exposed by the test_modeling_squeezebert.py
      
      fix an issue exposed by test_tokenization_squeezebert.py
      
      fix issue exposed by test_modeling_squeezebert.py
      
      auto generated code style improvement
      
      issue that we inherited from modeling_xxx.py: SqueezeBertForMaskedLM.forward() calls self.cls(), but there is no self.cls, and I think the goal was actually to call self.lm_head()
      
      update copyright
      
      resolve failing 'test_hidden_states_output' and remove unused encoder_hidden_states and encoder_attention_mask
      
      docs
      
      add integration test. rename squeezebert-mnli --> squeezebert/squeezebert-mnli
      
      autogenerated formatting tweaks
      
      integrate feedback from patrickvonplaten and sgugger to programming style and documentation strings
      
      * tiny change to order of imports
      02ef825b
  10. 01 Oct, 2020 9 commits