1. 18 Apr, 2024 1 commit
  2. 12 Feb, 2024 1 commit
  3. 02 Feb, 2024 1 commit
    • Klaus Hipp's avatar
      [Docs] Fix spelling and grammar mistakes (#28825) · 721ee783
      Klaus Hipp authored
      * Fix typos and grammar mistakes in docs and examples
      
      * Fix typos in docstrings and comments
      
      * Fix spelling of `tokenizer` in model tests
      
      * Remove erroneous spaces in decorators
      
      * Remove extra spaces in Markdown link texts
      721ee783
  4. 12 Oct, 2023 1 commit
  5. 05 Sep, 2023 1 commit
  6. 25 Aug, 2023 1 commit
  7. 25 Apr, 2023 1 commit
  8. 22 Feb, 2023 1 commit
  9. 06 Feb, 2023 1 commit
    • Sylvain Gugger's avatar
      Update quality tooling for formatting (#21480) · 6f79d264
      Sylvain Gugger authored
      * Result of black 23.1
      
      * Update target to Python 3.7
      
      * Switch flake8 to ruff
      
      * Configure isort
      
      * Configure isort
      
      * Apply isort with line limit
      
      * Put the right black version
      
      * adapt black in check copies
      
      * Fix copies
      6f79d264
  10. 27 Sep, 2022 1 commit
  11. 03 Aug, 2022 1 commit
    • LSinev's avatar
      Fix torch version comparisons (#18460) · 02b176c4
      LSinev authored
      Comparisons like
      version.parse(torch.__version__) > version.parse("1.6")
      are True for torch==1.6.0+cu101 or torch==1.6.0+cpu
      
      version.parse(version.parse(torch.__version__).base_version) are preferred (and available in pytorch_utils.py
      02b176c4
  12. 29 Jul, 2022 1 commit
  13. 19 May, 2022 1 commit
  14. 12 May, 2022 1 commit
  15. 30 Mar, 2022 1 commit
  16. 23 Mar, 2022 1 commit
  17. 12 Mar, 2022 1 commit
    • Stas Bekman's avatar
      [Deepspeed] add support for bf16 mode (#14569) · 580dd87c
      Stas Bekman authored
      
      
      * [WIP] add support for bf16 mode
      
      * prep for bf16
      
      * prep for bf16
      
      * fix; zero2/bf16 is ok
      
      * check bf16 is available
      
      * test fixes
      
      * enable zero3_bf16
      
      * config files
      
      * docs
      
      * split stage_dtype; merge back to non-dtype-specific config file
      
      * fix doc
      
      * cleanup
      
      * cleanup
      
      * bfloat16 => bf16 to match the PR changes
      
      * s/zero_gather_fp16_weights_on_model_save/zero_gather_16bit_weights_on_model_save/; s/save_fp16_model/save_16bit_model/
      
      * test fixes/skipping
      
      * move
      
      * fix
      
      * Update docs/source/main_classes/deepspeed.mdx
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * backticks
      
      * cleanup
      
      * cleanup
      
      * cleanup
      
      * new version
      
      * add note about grad accum in bf16
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      580dd87c
  18. 09 Feb, 2022 1 commit
  19. 10 Jan, 2022 1 commit
  20. 11 Nov, 2021 1 commit
  21. 22 Oct, 2021 1 commit
  22. 14 Oct, 2021 1 commit
  23. 08 Aug, 2021 1 commit
  24. 30 Jul, 2021 1 commit
    • 21jun's avatar
      fix typo in gradient_checkpointing arg (#12855) · 5c673efa
      21jun authored
      help for `ModelArguments.gradient_checkpointing` should be
      "If True, use gradient checkpointing to save memory
      at the expense of slower backward pass."
      not "Whether to freeze the feature extractor layers of the model."
      (which is duplicated from `freeze_feature_extractor` arg)
      5c673efa
  25. 23 Jul, 2021 1 commit
  26. 15 Jul, 2021 1 commit
  27. 25 Jun, 2021 1 commit
  28. 14 Jun, 2021 1 commit
  29. 09 Jun, 2021 2 commits
  30. 08 Jun, 2021 1 commit
  31. 12 May, 2021 1 commit
  32. 14 Apr, 2021 1 commit
  33. 30 Mar, 2021 1 commit
  34. 22 Mar, 2021 2 commits
  35. 21 Mar, 2021 4 commits