1. 13 May, 2022 1 commit
    • Yih-Dar's avatar
      Update self-push workflow (#17177) · 38043d84
      Yih-Dar authored
      
      
      * update push ci
      
      * install git-python
      
      * update comment
      
      * update deepspeed jobs
      
      * fix report
      
      * skip 2 more tests that require fairscale
      
      * Fix changes in test_fetcher.py (to deal with `setup.py` is changed)
      
      * set RUN_PT_TF_CROSS_TESTS=1 and final clean-up
      
      * remove SIGOPT_API_TOKEN
      
      * remove echo "$matrix_folders"
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      38043d84
  2. 12 May, 2022 1 commit
  3. 25 Apr, 2022 1 commit
  4. 19 Apr, 2022 1 commit
  5. 23 Mar, 2022 1 commit
    • Sylvain Gugger's avatar
      Reorganize file utils (#16264) · 4975002d
      Sylvain Gugger authored
      * Split file_utils in several submodules
      
      * Fixes
      
      * Add back more objects
      
      * More fixes
      
      * Who exactly decided to import that from there?
      
      * Second suggestion to code with code review
      
      * Revert wront move
      
      * Fix imports
      
      * Adapt all imports
      
      * Adapt all imports everywhere
      
      * Revert this import, will fix in a separate commit
      4975002d
  6. 23 Dec, 2021 1 commit
  7. 20 Jul, 2021 1 commit
  8. 14 Jul, 2021 1 commit
  9. 12 Jul, 2021 1 commit
  10. 22 Jun, 2021 1 commit
  11. 15 Jun, 2021 1 commit
  12. 06 May, 2021 1 commit
  13. 26 Apr, 2021 1 commit
  14. 21 Apr, 2021 1 commit
  15. 08 Apr, 2021 1 commit
  16. 15 Mar, 2021 1 commit
  17. 09 Mar, 2021 1 commit
  18. 08 Mar, 2021 1 commit
  19. 25 Feb, 2021 1 commit
  20. 15 Feb, 2021 1 commit
  21. 08 Feb, 2021 1 commit
  22. 15 Jan, 2021 1 commit
  23. 14 Jan, 2021 1 commit
  24. 13 Jan, 2021 1 commit
    • Stas Bekman's avatar
      [trainer] deepspeed integration (#9211) · 2df34f4a
      Stas Bekman authored
      
      
      * deepspeed integration
      
      * style
      
      * add test
      
      * ds wants to do its own backward
      
      * fp16 assert
      
      * Update src/transformers/training_args.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * style
      
      * for clarity extract what args are being passed to deepspeed
      
      * introduce the concept of self.wrapped_model
      
      * s/self.wrapped_model/self.model_wrapped/
      
      * complete transition to self.wrapped_model / self.model
      
      * fix
      
      * doc
      
      * give ds its own init
      
      * add custom overrides, handle bs correctly
      
      * fix test
      
      * clean up model_init logic, fix small bug
      
      * complete fix
      
      * collapse --deepspeed_config into --deepspeed
      
      * style
      
      * start adding doc notes
      
      * style
      
      * implement hf2ds optimizer and scheduler configuration remapping
      
      * oops
      
      * call get_num_training_steps absolutely when needed
      
      * workaround broken auto-formatter
      
      * deepspeed_config arg is no longer needed - fixed in deepspeed master
      
      * use hf's fp16 args in config
      
      * clean
      
      * start on the docs
      
      * rebase cleanup
      
      * finish up --fp16
      
      * clarify the supported stages
      
      * big refactor thanks to discovering deepspeed.init_distributed
      
      * cleanup
      
      * revert fp16 part
      
      * add checkpoint-support
      
      * more init ds into integrations
      
      * extend docs
      
      * cleanup
      
      * unfix docs
      
      * clean up old code
      
      * imports
      
      * move docs
      
      * fix logic
      
      * make it clear which file it's referring to
      
      * document nodes/gpus
      
      * style
      
      * wrong format
      
      * style
      
      * deepspeed handles gradient clipping
      
      * easier to read
      
      * major doc rewrite
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * docs
      
      * switch to AdamW optimizer
      
      * style
      
      * Apply suggestions from code review
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * clarify doc
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      2df34f4a
  25. 23 Dec, 2020 1 commit
  26. 22 Dec, 2020 2 commits
  27. 18 Dec, 2020 1 commit
  28. 17 Dec, 2020 1 commit
  29. 11 Dec, 2020 1 commit
  30. 30 Nov, 2020 1 commit
    • Stas Bekman's avatar
      [s2s trainer] fix DP mode (#8823) · 7f34d757
      Stas Bekman authored
      * fix DP case on multi-gpu
      
      * make executable
      
      * test all 3 modes
      
      * use the correct check for distributed
      
      * dp doesn't need a special case
      
      * restore original name
      
      * cleanup
      7f34d757
  31. 23 Nov, 2020 1 commit
  32. 18 Nov, 2020 1 commit
  33. 17 Nov, 2020 1 commit
  34. 08 Nov, 2020 1 commit
  35. 05 Nov, 2020 1 commit
  36. 28 Oct, 2020 1 commit
  37. 26 Oct, 2020 1 commit
  38. 23 Oct, 2020 1 commit
  39. 22 Oct, 2020 1 commit