1. 11 Mar, 2024 2 commits
    • Yitong Huang's avatar
      Make torch xla available on GPU (#29334) · 873d9bb3
      Yitong Huang authored
      
      
      * add USE_TORCH_XLA env
      
      * rename torch_tpu to torch_xla
      
      * better is_torch_xla_available; fix some fsdp and performance issues
      
      * fix format
      
      * fix bug when pjrt_device is cpu
      
      * fix bug
      
      * fix the deprecation handling
      
      ---------
      Co-authored-by: default avataranw90 <ang868@gmail.com>
      Co-authored-by: default avatarwangang.wa <wangang.wa@alibaba-inc.com>
      873d9bb3
    • Tanay Mehta's avatar
      Add Fill-in-the-middle training objective example - PyTorch (#27464) · 6d67837f
      Tanay Mehta authored
      * add: initial script to train clm fim
      
      * fix: if training model from scratch, new tokens will be added and embeddings resized
      
      * fix: fixed attention_mask errors when generating FIM data
      
      * fix: file formatted using black
      
      * add: run_fim_no_trainer.py and fixed some comments in run_fim.py
      
      * add: added fim examples to the README.md and ran code fixup
      
      * fix: little bug in both fim training scripts
      
      * fix: remove comment from notebook and added a note on fim related params
      
      * fix: minor typo in README
      
      * add: suggested minor changes to README and run_fim.py
      
      * add: gradient_accumulation_steps and gradient_checkpointing args
      
      * add: improved model embedding resizing
      
      * add: pad_to_multiple_of and attn_implementation params
      
      * add: requested minor changes
      
      * add: deepspeed zero compatibility
      
      * add: resize embeddings layer with zero3 support for fim model initialization
      6d67837f
  2. 21 Feb, 2024 1 commit
  3. 16 Feb, 2024 1 commit
  4. 01 Feb, 2024 1 commit
  5. 29 Jan, 2024 1 commit
  6. 26 Jan, 2024 1 commit
  7. 22 Jan, 2024 1 commit
  8. 19 Jan, 2024 1 commit
  9. 11 Jan, 2024 1 commit
    • Alex Hedges's avatar
      Set `cache_dir` for `evaluate.load()` in example scripts (#28422) · 95091e15
      Alex Hedges authored
      While using `run_clm.py`,[^1] I noticed that some files were being added
      to my global cache, not the local cache. I set the `cache_dir` parameter
      for the one call to `evaluate.load()`, which partially solved the
      problem. I figured that while I was fixing the one script upstream, I
      might as well fix the problem in all other example scripts that I could.
      
      There are still some files being added to my global cache, but this
      appears to be a bug in `evaluate` itself. This commit at least moves
      some of the files into the local cache, which is better than before.
      
      To create this PR, I made the following regex-based transformation:
      `evaluate\.load\((.*?)\)` -> `evaluate\.load\($1,
      cache_dir=model_args.cache_dir\)`. After using that, I manually fixed
      all modified files with `ruff` serving as useful guidance. During the
      process, I removed one existing usage of the `cache_dir` parameter in a
      script that did not have a corresponding `--cache-dir` argument
      declared.
      
      [^1]: I specifically used `pytorch/language-modeling/run_clm.py` from
      v4.34.1 of the library. For the original code, see the following URL:
      https://github.com/huggingface/transformers/tree/acc394c4f5e1283c19783581790b3dc3105a3697/examples/pytorch/language-modeling/run_clm.py.
      95091e15
  10. 13 Dec, 2023 1 commit
  11. 11 Dec, 2023 1 commit
    • Adam Louly's avatar
      fix no sequence length models error (#27522) · 4850aaba
      Adam Louly authored
      * fix no sequence length models error
      
      * block size check
      
      ---------
      
      Co-authored-by: Adam Louly <adamlouly@microsoft.com@orttrainingdev9.d32nl1ml4oruzj4qz3bqlggovf.px.internal.cloudapp.net>
      4850aaba
  12. 17 Nov, 2023 1 commit
  13. 15 Nov, 2023 1 commit
  14. 02 Nov, 2023 1 commit
  15. 31 Oct, 2023 1 commit
  16. 27 Oct, 2023 1 commit
  17. 12 Oct, 2023 1 commit
  18. 11 Oct, 2023 1 commit
  19. 04 Oct, 2023 1 commit
    • Phuc Van Phan's avatar
      refactor: change default block_size (#26229) · 6015f91a
      Phuc Van Phan authored
      * refactor: change default block_size
      
      * fix: return tf to origin
      
      * fix: change files to origin
      
      * rebase
      
      * rebase
      
      * rebase
      
      * rebase
      
      * rebase
      
      * rebase
      
      * rebase
      
      * rebase
      
      * refactor: add min block_size to files
      
      * reformat: add min block_size for run_clm tf
      6015f91a
  20. 03 Oct, 2023 1 commit
  21. 28 Sep, 2023 1 commit
  22. 12 Sep, 2023 1 commit
  23. 11 Sep, 2023 2 commits
  24. 04 Sep, 2023 1 commit
  25. 23 Aug, 2023 1 commit
  26. 21 Aug, 2023 1 commit
  27. 08 Aug, 2023 1 commit
  28. 07 Aug, 2023 2 commits
  29. 02 Aug, 2023 1 commit
  30. 28 Jul, 2023 2 commits
  31. 20 Jul, 2023 1 commit
  32. 17 Jul, 2023 1 commit
  33. 12 Jun, 2023 1 commit
  34. 07 Jun, 2023 1 commit
  35. 06 Jun, 2023 1 commit
  36. 22 May, 2023 1 commit