1. 24 Aug, 2020 4 commits
  2. 20 Aug, 2020 4 commits
  3. 14 Aug, 2020 1 commit
    • Jin Young (Daniel) Sohn's avatar
      Fix TPU Convergence bug introduced by PR#6151 (#6488) · 24107c2c
      Jin Young (Daniel) Sohn authored
      Currently with the bug introduced we're taking two optimizer steps per
      batch: one global one, where `xm.optimizer_step` injects a CRS between
      all cores in training, and one without. This has been affecting training
      accuracy (for example, XLNet GLUE on MNLI is not converging, etc.).
      24107c2c
  4. 12 Aug, 2020 1 commit
  5. 11 Aug, 2020 1 commit
  6. 06 Aug, 2020 1 commit
  7. 05 Aug, 2020 1 commit
  8. 03 Aug, 2020 2 commits
  9. 31 Jul, 2020 2 commits
  10. 30 Jul, 2020 1 commit
    • Sylvain Gugger's avatar
      Switch from return_tuple to return_dict (#6138) · 91cb9546
      Sylvain Gugger authored
      
      
      * Switch from return_tuple to return_dict
      
      * Fix test
      
      * [WIP] Test TF Flaubert + Add {XLM, Flaubert}{TokenClassification, MultipleC… (#5614)
      
      * Test TF Flaubert + Add {XLM, Flaubert}{TokenClassification, MultipleChoice} models and tests
      
      * AutoModels
      
      
      Tiny tweaks
      
      * Style
      
      * Final changes before merge
      
      * Re-order for simpler review
      
      * Final fixes
      
      * Addressing @sgugger's comments
      
      * Test MultipleChoice
      
      * Rework TF trainer (#6038)
      
      * Fully rework training/prediction loops
      
      * fix method name
      
      * Fix variable name
      
      * Fix property name
      
      * Fix scope
      
      * Fix method name
      
      * Fix tuple index
      
      * Fix tuple index
      
      * Fix indentation
      
      * Fix variable name
      
      * fix eval before log
      
      * Add drop remainder for test dataset
      
      * Fix step number + fix logging datetime
      
      * fix eval loss value
      
      * use global step instead of step + fix logging at step 0
      
      * Fix logging datetime
      
      * Fix global_step usage
      
      * Fix breaking loop + logging datetime
      
      * Fix step in prediction loop
      
      * Fix step breaking
      
      * Fix train/test loops
      
      * Force TF at least 2.2 for the trainer
      
      * Use assert_cardinality to facilitate the dataset size computation
      
      * Log steps per epoch
      
      * Make tfds compliant with TPU
      
      * Make tfds compliant with TPU
      
      * Use TF dataset enumerate instead of the Python one
      
      * revert previous commit
      
      * Fix data_dir
      
      * Apply style
      
      * rebase on master
      
      * Address Sylvain's comments
      
      * Address Sylvain's and Lysandre comments
      
      * Trigger CI
      
      * Remove unused import
      
      * Switch from return_tuple to return_dict
      
      * Fix test
      
      * Add recent model
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      Co-authored-by: default avatarJulien Plu <plu.julien@gmail.com>
      91cb9546
  11. 28 Jul, 2020 1 commit
  12. 27 Jul, 2020 1 commit
  13. 26 Jul, 2020 1 commit
  14. 23 Jul, 2020 1 commit
  15. 20 Jul, 2020 3 commits
    • Sylvain Gugger's avatar
      Clarify arg class (#5916) · 4781afd0
      Sylvain Gugger authored
      4781afd0
    • Stas Bekman's avatar
      DataParallel fixes (#5733) · 35cb101e
      Stas Bekman authored
      * DataParallel fixes:
      
      1. switched to a more precise check
      -        if self.args.n_gpu > 1:
      +        if isinstance(model, nn.DataParallel):
      
      2. fix tests - require the same fixup under DataParallel as the training module
      
      * another fix
      35cb101e
    • Pradhy729's avatar
      Trainer support for iterabledataset (#5834) · 290b6e18
      Pradhy729 authored
      * Don't pass sampler for iterable dataset
      
      * Added check for test and eval dataloaders.
      
      * Formatting
      
      * Don't pass sampler for iterable dataset
      
      * Added check for test and eval dataloaders.
      
      * Formatting
      
      * Cleaner if nesting.
      
      * Added test for trainer and iterable dataset
      
      * Formatting for test
      
      * Fixed import when torch is available only.
      
      * Added require torch decorator to helper class
      
      * Moved dataset class inside unittest
      
      * Removed nested if and changed model in test
      
      * Checking torch availability for IterableDataset
      290b6e18
  16. 13 Jul, 2020 1 commit
  17. 01 Jul, 2020 2 commits
  18. 30 Jun, 2020 1 commit
  19. 23 Jun, 2020 1 commit
  20. 22 Jun, 2020 1 commit
    • Patrick von Platen's avatar
      Benchmarks (#4912) · fa0be6d7
      Patrick von Platen authored
      * finish benchmark
      
      * fix isort
      
      * fix setup cfg
      
      * retab
      
      * fix time measuring of tf graph mode
      
      * fix tf cuda
      
      * clean code
      
      * better error message
      fa0be6d7
  21. 17 Jun, 2020 2 commits
  22. 16 Jun, 2020 1 commit
  23. 15 Jun, 2020 2 commits
  24. 11 Jun, 2020 1 commit
  25. 10 Jun, 2020 2 commits
  26. 09 Jun, 2020 1 commit