1. 18 Apr, 2024 1 commit
  2. 25 Mar, 2024 1 commit
  3. 04 Mar, 2024 1 commit
    • flybird11111's avatar
      [example]add gpt2 benchmark example script. (#5295) · 29695cf7
      flybird11111 authored
      
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed ddp test (#5254)
      
      * [ci] fixed ddp test
      
      * polish
      
      * fix typo in  applications/ColossalEval/README.md (#5250)
      
      * [ci] fix shardformer tests. (#5255)
      
      * fix ci
      
      fix
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [doc] fix doc typo (#5256)
      
      * [doc] fix annotation display
      
      * [doc] fix llama2 doc
      
      * [hotfix]: add pp sanity check and fix mbs arg (#5268)
      
      * fix: fix misleading mbs arg
      
      * feat: add pp sanity check
      
      * fix: fix 1f1b sanity check
      
      * [workflow] fixed incomplete bash command (#5272)
      
      * [workflow] fixed oom tests (#5275)
      
      * [workflow] fixed oom tests
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
      
      * fix ci
      
      fix
      
      * fix test
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      * fix
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
      
      * support gradients acc
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
      
      * fix auto loading gpt2 tokenizer (#5279)
      
      * [doc] add llama2-13B disyplay (#5285)
      
      * Update README.md
      
      * fix 13b typo
      
      ---------
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      
      * fix llama pretrain (#5287)
      
      * fix
      
      * fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      * Update shardformer.py
      
      ---------
      Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
      Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
      Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
      Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
      29695cf7
  4. 09 Jan, 2024 1 commit
  5. 21 Sep, 2023 1 commit
  6. 20 Sep, 2023 1 commit
    • Wenhao Chen's avatar
      [chat]: update rm, add wandb and fix bugs (#4471) · 7b9b8644
      Wenhao Chen authored
      
      
      * feat: modify forward fn of critic and reward model
      
      * feat: modify calc_action_log_probs
      
      * to: add wandb in sft and rm trainer
      
      * feat: update train_sft
      
      * feat: update train_rm
      
      * style: modify type annotation and add warning
      
      * feat: pass tokenizer to ppo trainer
      
      * to: modify trainer base and maker base
      
      * feat: add wandb in ppo trainer
      
      * feat: pass tokenizer to generate
      
      * test: update generate fn tests
      
      * test: update train tests
      
      * fix: remove action_mask
      
      * feat: remove unused code
      
      * fix: fix wrong ignore_index
      
      * fix: fix mock tokenizer
      
      * chore: update requirements
      
      * revert: modify make_experience
      
      * fix: fix inference
      
      * fix: add padding side
      
      * style: modify _on_learn_batch_end
      
      * test: use mock tokenizer
      
      * fix: use bf16 to avoid overflow
      
      * fix: fix workflow
      
      * [chat] fix gemini strategy
      
      * [chat] fix
      
      * sync: update colossalai strategy
      
      * fix: fix args and model dtype
      
      * fix: fix checkpoint test
      
      * fix: fix requirements
      
      * fix: fix missing import and wrong arg
      
      * fix: temporarily skip gemini test in stage 3
      
      * style: apply pre-commit
      
      * fix: temporarily skip gemini test in stage 1&2
      
      ---------
      Co-authored-by: default avatarMingyan Jiang <1829166702@qq.com>
      7b9b8644
  7. 19 Sep, 2023 1 commit
  8. 18 Sep, 2023 2 commits
    • github-actions[bot]'s avatar
    • Hongxin Liu's avatar
      [legacy] clean up legacy code (#4743) · b5f9e37c
      Hongxin Liu authored
      * [legacy] remove outdated codes of pipeline (#4692)
      
      * [legacy] remove cli of benchmark and update optim (#4690)
      
      * [legacy] remove cli of benchmark and update optim
      
      * [doc] fix cli doc test
      
      * [legacy] fix engine clip grad norm
      
      * [legacy] remove outdated colo tensor (#4694)
      
      * [legacy] remove outdated colo tensor
      
      * [test] fix test import
      
      * [legacy] move outdated zero to legacy (#4696)
      
      * [legacy] clean up utils (#4700)
      
      * [legacy] clean up utils
      
      * [example] update examples
      
      * [legacy] clean up amp
      
      * [legacy] fix amp module
      
      * [legacy] clean up gpc (#4742)
      
      * [legacy] clean up context
      
      * [legacy] clean core, constants and global vars
      
      * [legacy] refactor initialize
      
      * [example] fix examples ci
      
      * [example] fix examples ci
      
      * [legacy] fix tests
      
      * [example] fix gpt example
      
      * [example] fix examples ci
      
      * [devops] fix ci installation
      
      * [example] fix examples ci
      b5f9e37c
  9. 15 Sep, 2023 1 commit
  10. 11 Sep, 2023 1 commit
    • Hongxin Liu's avatar
      [legacy] move communication and nn to legacy and refactor logger (#4671) · 554aa959
      Hongxin Liu authored
      * [legacy] move communication to legacy (#4640)
      
      * [legacy] refactor logger and clean up legacy codes (#4654)
      
      * [legacy] make logger independent to gpc
      
      * [legacy] make optim independent to registry
      
      * [legacy] move test engine to legacy
      
      * [legacy] move nn to legacy (#4656)
      
      * [legacy] move nn to legacy
      
      * [checkpointio] fix save hf config
      
      * [test] remove useledd rpc pp test
      
      * [legacy] fix nn init
      
      * [example] skip tutorial hybriad parallel example
      
      * [devops] test doc check
      
      * [devops] test doc check
      554aa959
  11. 05 Sep, 2023 2 commits
  12. 24 Aug, 2023 1 commit
    • Hongxin Liu's avatar
      [gemini] improve compatibility and add static placement policy (#4479) · 27061426
      Hongxin Liu authored
      * [gemini] remove distributed-related part from colotensor (#4379)
      
      * [gemini] remove process group dependency
      
      * [gemini] remove tp part from colo tensor
      
      * [gemini] patch inplace op
      
      * [gemini] fix param op hook and update tests
      
      * [test] remove useless tests
      
      * [test] remove useless tests
      
      * [misc] fix requirements
      
      * [test] fix model zoo
      
      * [test] fix model zoo
      
      * [test] fix model zoo
      
      * [test] fix model zoo
      
      * [test] fix model zoo
      
      * [misc] update requirements
      
      * [gemini] refactor gemini optimizer and gemini ddp (#4398)
      
      * [gemini] update optimizer interface
      
      * [gemini] renaming gemini optimizer
      
      * [gemini] refactor gemini ddp class
      
      * [example] update gemini related example
      
      * [example] update gemini related example
      
      * [plugin] fix gemini plugin args
      
      * [test] update gemini ckpt tests
      
      * [gemini] fix checkpoint io
      
      * [example] fix opt example requirements
      
      * [example] fix opt example
      
      * [example] fix opt example
      
      * [example] fix opt example
      
      * [gemini] add static placement policy (#4443)
      
      * [gemini] add static placement policy
      
      * [gemini] fix param offload
      
      * [test] update gemini tests
      
      * [plugin] update gemini plugin
      
      * [plugin] update gemini plugin docstr
      
      * [misc] fix flash attn requirement
      
      * [test] fix gemini checkpoint io test
      
      * [example] update resnet example result (#4457)
      
      * [example] update bert example result (#4458)
      
      * [doc] update gemini doc (#4468)
      
      * [example] update gemini related examples (#4473)
      
      * [example] update gpt example
      
      * [example] update dreambooth example
      
      * [example] update vit
      
      * [example] update opt
      
      * [example] update palm
      
      * [example] update vit and opt benchmark
      
      * [hotfix] fix bert in model zoo (#4480)
      
      * [hotfix] fix bert in model zoo
      
      * [test] remove chatglm gemini test
      
      * [test] remove sam gemini test
      
      * [test] remove vit gemini test
      
      * [hotfix] fix opt tutorial example (#4497)
      
      * [hotfix] fix opt tutorial example
      
      * [hotfix] fix opt tutorial example
      27061426
  13. 28 Jun, 2023 1 commit
  14. 26 Jun, 2023 1 commit
  15. 19 Jun, 2023 1 commit
  16. 08 Jun, 2023 1 commit
  17. 30 May, 2023 1 commit
  18. 18 May, 2023 1 commit
  19. 26 Apr, 2023 1 commit
    • digger-yu's avatar
      [doc] Fix typo under colossalai and doc(#3618) · b9a8dff7
      digger-yu authored
      * Fixed several spelling errors under colossalai
      
      * Fix the spelling error in colossalai and docs directory
      
      * Cautious Changed the spelling error under the example folder
      
      * Update runtime_preparation_pass.py
      
      revert autograft to autograd
      
      * Update search_chunk.py
      
      utile to until
      
      * Update check_installation.py
      
      change misteach to mismatch in line 91
      
      * Update 1D_tensor_parallel.md
      
      revert to perceptron
      
      * Update 2D_tensor_parallel.md
      
      revert to perceptron in line 73
      
      * Update 2p5D_tensor_parallel.md
      
      revert to perceptron in line 71
      
      * Update 3D_tensor_parallel.md
      
      revert to perceptron in line 80
      
      * Update README.md
      
      revert to resnet in line 42
      
      * Update reorder_graph.py
      
      revert to indice in line 7
      
      * Update p2p.py
      
      revert to megatron in line 94
      
      * Update initialize.py
      
      revert to torchrun in line 198
      
      * Update routers.py
      
      change to detailed in line 63
      
      * Update routers.py
      
      change to detailed in line 146
      
      * Update README.md
      
      revert  random number in line 402
      b9a8dff7
  20. 06 Apr, 2023 1 commit
  21. 04 Apr, 2023 1 commit
    • ver217's avatar
      [zero] reorganize zero/gemini folder structure (#3424) · 26b7aac0
      ver217 authored
      * [zero] refactor low-level zero folder structure
      
      * [zero] fix legacy zero import path
      
      * [zero] fix legacy zero import path
      
      * [zero] remove useless import
      
      * [zero] refactor gemini folder structure
      
      * [zero] refactor gemini folder structure
      
      * [zero] refactor legacy zero import path
      
      * [zero] refactor gemini folder structure
      
      * [zero] refactor gemini folder structure
      
      * [zero] refactor gemini folder structure
      
      * [zero] refactor legacy zero import path
      
      * [zero] fix test import path
      
      * [zero] fix test
      
      * [zero] fix circular import
      
      * [zero] update import
      26b7aac0
  22. 23 Mar, 2023 1 commit
  23. 21 Mar, 2023 1 commit
  24. 07 Mar, 2023 1 commit
  25. 27 Feb, 2023 2 commits
  26. 22 Feb, 2023 1 commit
  27. 15 Feb, 2023 1 commit
  28. 09 Feb, 2023 1 commit
  29. 31 Jan, 2023 1 commit
  30. 30 Jan, 2023 1 commit
  31. 28 Jan, 2023 1 commit
  32. 20 Jan, 2023 1 commit
  33. 18 Jan, 2023 3 commits
  34. 17 Jan, 2023 1 commit
  35. 16 Jan, 2023 1 commit