"...resnet50_tensorflow.git" did not exist on "0b395f650aa8c95afb219553e8ca654b493660e3"
  1. 07 May, 2020 1 commit
    • Julien Chaumond's avatar
      BIG Reorganize examples (#4213) · 0ae96ff8
      Julien Chaumond authored
      * Created using Colaboratory
      
      * [examples] reorganize files
      
      * remove run_tpu_glue.py as superseded by TPU support in Trainer
      
      * Bugfix: int, not tuple
      
      * move files around
      0ae96ff8
  2. 06 May, 2020 1 commit
    • Julien Plu's avatar
      TF version of the trainer (#4017) · aad50151
      Julien Plu authored
      * First commit to add a TF version of the trainer.
      
      * Make the TF trainer closer to what looks the PT trainer
      
      * Refactoring common code between the PT and TF trainer into an util file.
      
      * Some bugfix + better similarity with the PT trainer
      
      * Add missing class in transformers init
      
      * Bugfix over prediction + use classification report instead of simple metrics
      
      * Fix name error
      
      * Fix optimization tests + style
      
      * Apply style
      
      * Several bugfix for multi-gpu training
      
      * Apply style
      
      * Apply style
      
      * Add glue example for the TF trainer
      
      * Several bugix + address the reviews
      
      * Fix on the TF training args file
      
      * Add a debug mode
      
      * Bugfix in utils_ner.py when segment_ids is None
      
      * Apply style
      
      * Apply style
      
      * Add TPU strategy
      
      * Fix selection strategy
      aad50151
  3. 22 Apr, 2020 1 commit
    • Julien Chaumond's avatar
      Trainer (#3800) · dd9d483d
      Julien Chaumond authored
      * doc
      
      * [tests] Add sample files for a regression task
      
      * [HUGE] Trainer
      
      * Feedback from @sshleifer
      
      * Feedback from @thomwolf + logging tweak
      
      * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes
      
      * [glue] Use default max_seq_length of 128 like before
      
      * [glue] move DataTrainingArguments around
      
      * [ner] Change interface of InputExample, and align run_{tf,pl}
      
      * Re-align the pl scripts a little bit
      
      * ner
      
      * [ner] Add integration test
      
      * Fix language_modeling with API tweak
      
      * [ci] Tweak loss target
      
      * Don't break console output
      
      * amp.initialize: model must be on right device before
      
      * [multiple-choice] update for Trainer
      
      * Re-align to 827d6d6e
      dd9d483d
  4. 27 Mar, 2020 1 commit
  5. 20 Feb, 2020 1 commit
  6. 01 Feb, 2020 1 commit
  7. 06 Jan, 2020 2 commits
  8. 22 Dec, 2019 3 commits
  9. 21 Dec, 2019 1 commit
    • Aymeric Augustin's avatar
      Reformat source code with black. · fa84ae26
      Aymeric Augustin authored
      This is the result of:
      
          $ black --line-length 119 examples templates transformers utils hubconf.py setup.py
      
      There's a lot of fairly long lines in the project. As a consequence, I'm
      picking the longest widely accepted line length, 119 characters.
      
      This is also Thomas' preference, because it allows for explicit variable
      names, to make the code easier to understand.
      fa84ae26
  10. 12 Dec, 2019 1 commit
  11. 15 Oct, 2019 8 commits