1. 26 Apr, 2021 1 commit
  2. 23 Apr, 2021 2 commits
  3. 22 Apr, 2021 1 commit
  4. 21 Apr, 2021 1 commit
  5. 20 Apr, 2021 2 commits
  6. 19 Apr, 2021 2 commits
  7. 16 Apr, 2021 1 commit
  8. 15 Apr, 2021 1 commit
  9. 14 Apr, 2021 1 commit
  10. 08 Apr, 2021 4 commits
  11. 31 Mar, 2021 2 commits
  12. 29 Mar, 2021 1 commit
    • pcuenca's avatar
      Allow use of pre-computed lengths when grouping by length. (#10953) · ae6b6963
      pcuenca authored
      A new argument `length_column_name` has been added to
      `TrainingArguments`, with default value `"length"`. If this column
      exists and `group_by_length` is `True`, the train sampler will use
      it for grouping rather than computing it before training starts.
      
      This is an optimization that allows the user to prepare data for fast
      processing, preventing sequential access to the dataset as described in
      issue #10909.
      ae6b6963
  13. 24 Mar, 2021 1 commit
  14. 23 Mar, 2021 1 commit
  15. 22 Mar, 2021 2 commits
  16. 18 Mar, 2021 1 commit
  17. 17 Mar, 2021 3 commits
    • Mansi Mane's avatar
      Smmp batch not divisible by microbatches fix (#10778) · 0282e24e
      Mansi Mane authored
      
      
      * Added debug prints
      
      * Added config
      
      * Added prints
      
      * Added prints
      
      * Added extra samples to SequentialDistributedSampler
      
      * Added extra samples to SequentialDistributedSampler
      
      Updated SequentialDistributedSampler call
      
      * Added deubg prints
      
      * Removed extra prints
      
      * Making predicitons and labels multiple of batchsize
      
      * updated number of microbatches
      
      * Removed extra prints
      
      * Made start_remainder similar to DistributedSamplerWithLoop
      
      * Minor spacing update
      
      * Added debug prints
      
      Added config
      
      Added prints
      
      Added prints
      
      * Added extra samples to SequentialDistributedSampler
      
      Updated SequentialDistributedSampler call
      
      Added extra samples to SequentialDistributedSampler
      
      Added deubg prints
      
      Removed extra prints
      
      Making predicitons and labels multiple of batchsize
      
      updated number of microbatches
      
      Removed extra prints
      
      Squashing redundant commits
      
      * Made start_remainder similar to DistributedSamplerWithLoop
      
      Minor spacing update
      
      Made start_remainder similar to DistributedSamplerWithLoop
      
      * Test and styling
      
      * Rename test
      Co-authored-by: default avatarSylvain Gugger <sylvain.gugger@gmail.com>
      0282e24e
    • Stas Bekman's avatar
    • Stas Bekman's avatar
      [DeepSpeed] improve checkpoint loading code plus tests (#10760) · cd8c93f7
      Stas Bekman authored
      * deepspeed checkpoint loading code plus tests
      
      * style
      
      * style
      cd8c93f7
  18. 16 Mar, 2021 2 commits
  19. 15 Mar, 2021 2 commits
  20. 12 Mar, 2021 1 commit
  21. 11 Mar, 2021 1 commit
  22. 10 Mar, 2021 1 commit
    • Philipp Schmid's avatar
      Extend trainer logging for sm (#10633) · 49c61a4a
      Philipp Schmid authored
      * renamed logging to hf_logging
      
      * changed logging from hf_logging to logging and loggin to native_logging
      
      * removed everything trying to fix import Trainer error
      
      * adding imports again
      
      * added custom add_handler function to logging.py
      
      * make style
      
      * added remove_handler
      
      * added another conditional to assert
      49c61a4a
  23. 09 Mar, 2021 1 commit
  24. 08 Mar, 2021 4 commits
  25. 05 Mar, 2021 1 commit