1. 27 May, 2020 1 commit
    • Hao Tan's avatar
      Add back --do_lower_case to uncased models (#4245) · a9aa7456
      Hao Tan authored
      The option `--do_lower_case` is currently required by the uncased models (i.e., bert-base-uncased, bert-large-uncased).
      
      Results:
      BERT-BASE without --do_lower_case:  'exact': 73.83, 'f1': 82.22
      BERT-BASE with --do_lower_case:  'exact': 81.02, 'f1': 88.34
      a9aa7456
  2. 25 May, 2020 1 commit
  3. 21 May, 2020 2 commits
  4. 19 May, 2020 2 commits
  5. 18 May, 2020 2 commits
  6. 15 May, 2020 3 commits
  7. 14 May, 2020 2 commits
  8. 13 May, 2020 2 commits
  9. 12 May, 2020 1 commit
    • Viktor Alm's avatar
      Add MultipleChoice to TFTrainer [WIP] (#4270) · e4512aab
      Viktor Alm authored
      
      
      * catch gpu len 1 set to gpu0
      
      * Add mpc to trainer
      
      * Add MPC for TF
      
      * fix TF automodel for MPC and add Albert
      
      * Apply style
      
      * Fix import
      
      * Note to self: double check
      
      * Make shape None, None for datasetgenerator output shapes
      
      * Add from_pt bool which doesnt seem to work
      
      * Original checkpoint dir
      
      * Fix docstrings for automodel
      
      * Update readme and apply style
      
      * Colab should probably not be from users
      
      * Colabs should probably not be from users
      
      * Add colab
      
      * Update README.md
      
      * Update README.md
      
      * Cleanup __intit__
      
      * Cleanup flake8 trailing comma
      
      * Update src/transformers/training_args_tf.py
      
      * Update src/transformers/modeling_tf_auto.py
      Co-authored-by: default avatarViktor Alm <viktoralm@pop-os.localdomain>
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      e4512aab
  10. 11 May, 2020 1 commit
  11. 08 May, 2020 1 commit
  12. 07 May, 2020 5 commits
  13. 06 May, 2020 2 commits
    • Julien Plu's avatar
      TF version of the trainer (#4017) · aad50151
      Julien Plu authored
      * First commit to add a TF version of the trainer.
      
      * Make the TF trainer closer to what looks the PT trainer
      
      * Refactoring common code between the PT and TF trainer into an util file.
      
      * Some bugfix + better similarity with the PT trainer
      
      * Add missing class in transformers init
      
      * Bugfix over prediction + use classification report instead of simple metrics
      
      * Fix name error
      
      * Fix optimization tests + style
      
      * Apply style
      
      * Several bugfix for multi-gpu training
      
      * Apply style
      
      * Apply style
      
      * Add glue example for the TF trainer
      
      * Several bugix + address the reviews
      
      * Fix on the TF training args file
      
      * Add a debug mode
      
      * Bugfix in utils_ner.py when segment_ids is None
      
      * Apply style
      
      * Apply style
      
      * Add TPU strategy
      
      * Fix selection strategy
      aad50151
    • Simone Primarosa's avatar
  14. 02 May, 2020 3 commits
  15. 01 May, 2020 1 commit
  16. 29 Apr, 2020 1 commit
    • Julien Chaumond's avatar
      CDN urls (#4030) · 455c6390
      Julien Chaumond authored
      * [file_utils] use_cdn + documentation
      
      * Move to cdn. urls for weights
      
      * [urls] Hotfix for bert-base-japanese
      455c6390
  17. 28 Apr, 2020 2 commits
  18. 24 Apr, 2020 2 commits
  19. 22 Apr, 2020 2 commits
    • Julien Chaumond's avatar
      Fixes #3877 · 1dc9b3c7
      Julien Chaumond authored
      1dc9b3c7
    • Julien Chaumond's avatar
      Trainer (#3800) · dd9d483d
      Julien Chaumond authored
      * doc
      
      * [tests] Add sample files for a regression task
      
      * [HUGE] Trainer
      
      * Feedback from @sshleifer
      
      * Feedback from @thomwolf + logging tweak
      
      * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes
      
      * [glue] Use default max_seq_length of 128 like before
      
      * [glue] move DataTrainingArguments around
      
      * [ner] Change interface of InputExample, and align run_{tf,pl}
      
      * Re-align the pl scripts a little bit
      
      * ner
      
      * [ner] Add integration test
      
      * Fix language_modeling with API tweak
      
      * [ci] Tweak loss target
      
      * Don't break console output
      
      * amp.initialize: model must be on right device before
      
      * [multiple-choice] update for Trainer
      
      * Re-align to 827d6d6e
      dd9d483d
  20. 20 Apr, 2020 3 commits
  21. 18 Apr, 2020 1 commit
    • Thomas Wolf's avatar
      Cleanup fast tokenizers integration (#3706) · 827d6d6e
      Thomas Wolf authored
      
      
      * First pass on utility classes and python tokenizers
      
      * finishing cleanup pass
      
      * style and quality
      
      * Fix tests
      
      * Updating following @mfuntowicz comment
      
      * style and quality
      
      * Fix Roberta
      
      * fix batch_size/seq_length inBatchEncoding
      
      * add alignement methods + tests
      
      * Fix OpenAI and Transfo-XL tokenizers
      
      * adding trim_offsets=True default for GPT2 et RoBERTa
      
      * style and quality
      
      * fix tests
      
      * add_prefix_space in roberta
      
      * bump up tokenizers to rc7
      
      * style
      
      * unfortunately tensorfow does like these - removing shape/seq_len for now
      
      * Update src/transformers/tokenization_utils.py
      Co-Authored-By: default avatarStefan Schweter <stefan@schweter.it>
      
      * Adding doc and docstrings
      
      * making flake8 happy
      Co-authored-by: default avatarStefan Schweter <stefan@schweter.it>
      827d6d6e