1. 18 Apr, 2024 1 commit
  2. 08 Apr, 2024 1 commit
  3. 07 Apr, 2024 2 commits
  4. 01 Apr, 2024 1 commit
    • Wenhao Chen's avatar
      [shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous... · e614aa34
      Wenhao Chen authored
      [shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508)
      
      * feat: add `GradientCheckpointConfig` and `PipelineGradientCheckpointConfig`
      
      * feat: apply `GradientCheckpointConfig` to policy and llama_forward
      
      * feat: move `distribute_layer` and `get_stage_index` to PipelineStageManager
      
      * fix: add optional args for `distribute_layer` and `get_stage_index`
      
      * fix: fix changed API calls
      
      * test: update llama tests
      
      * style: polish `GradientCheckpointConfig`
      
      * fix: fix pipeline utils tests
      e614aa34
  5. 28 Mar, 2024 1 commit
  6. 27 Mar, 2024 1 commit
  7. 26 Mar, 2024 1 commit
  8. 25 Mar, 2024 2 commits
  9. 24 Mar, 2024 1 commit
    • Yuanheng Zhao's avatar
      [example] update Grok-1 inference (#5495) · 5fcd7795
      Yuanheng Zhao authored
      * revise grok-1 example
      
      * remove unused arg in scripts
      
      * prevent re-installing torch
      
      * update readme
      
      * revert modifying colossalai requirements
      
      * add perf
      
      * trivial
      
      * add tokenizer url
      5fcd7795
  10. 22 Mar, 2024 1 commit
  11. 21 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [example] add grok-1 inference (#5485) · 848a574c
      Hongxin Liu authored
      * [misc] add submodule
      
      * remove submodule
      
      * [example] support grok-1 tp inference
      
      * [example] add grok-1 inference script
      
      * [example] refactor code
      
      * [example] add grok-1 readme
      
      * [exmaple] add test ci
      
      * [exmaple] update readme
      848a574c
  12. 05 Mar, 2024 2 commits
  13. 04 Mar, 2024 1 commit
    • flybird11111's avatar
      [example]add gpt2 benchmark example script. (#5295) · 29695cf7
      flybird11111 authored
      
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed ddp test (#5254)
      
      * [ci] fixed ddp test
      
      * polish
      
      * fix typo in  applications/ColossalEval/README.md (#5250)
      
      * [ci] fix shardformer tests. (#5255)
      
      * fix ci
      
      fix
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [doc] fix doc typo (#5256)
      
      * [doc] fix annotation display
      
      * [doc] fix llama2 doc
      
      * [hotfix]: add pp sanity check and fix mbs arg (#5268)
      
      * fix: fix misleading mbs arg
      
      * feat: add pp sanity check
      
      * fix: fix 1f1b sanity check
      
      * [workflow] fixed incomplete bash command (#5272)
      
      * [workflow] fixed oom tests (#5275)
      
      * [workflow] fixed oom tests
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
      
      * fix ci
      
      fix
      
      * fix test
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      * fix
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
      
      * support gradients acc
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
      
      * fix auto loading gpt2 tokenizer (#5279)
      
      * [doc] add llama2-13B disyplay (#5285)
      
      * Update README.md
      
      * fix 13b typo
      
      ---------
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      
      * fix llama pretrain (#5287)
      
      * fix
      
      * fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      * Update shardformer.py
      
      ---------
      Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
      Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
      Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
      Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
      29695cf7
  14. 27 Feb, 2024 1 commit
  15. 30 Jan, 2024 1 commit
  16. 19 Jan, 2024 1 commit
  17. 15 Jan, 2024 1 commit
  18. 11 Jan, 2024 1 commit
  19. 09 Jan, 2024 1 commit
  20. 08 Jan, 2024 1 commit
    • Xuanlei Zhao's avatar
      [npu] use extension for op builder (#5172) · dd2c28a3
      Xuanlei Zhao authored
      * update extension
      
      * update cpu adam
      
      * update is
      
      * add doc for cpu adam
      
      * update kernel
      
      * update commit
      
      * update flash
      
      * update memory efficient
      
      * update flash attn
      
      * update flash attention loader
      
      * update api
      
      * fix
      
      * update doc
      
      * update example time limit
      
      * reverse change
      
      * fix doc
      
      * remove useless kernel
      
      * fix
      
      * not use warning
      
      * update
      
      * update
      dd2c28a3
  21. 02 Jan, 2024 1 commit
  22. 22 Dec, 2023 1 commit
    • Wenhao Chen's avatar
      [pipeline]: fix p2p comm, add metadata cache and support llama interleaved pp (#5134) · 4fa689fc
      Wenhao Chen authored
      * test: add more p2p tests
      
      * fix: remove send_forward_recv_forward as p2p op list need to use the same group
      
      * fix: make send and receive atomic
      
      * feat: update P2PComm fn
      
      * feat: add metadata cache in 1f1b
      
      * feat: add metadata cache in interleaved pp
      
      * feat: modify is_xx_stage fn
      
      * revert: add _broadcast_object_list
      
      * feat: add interleaved pp in llama policy
      
      * feat: set NCCL_BUFFSIZE in HybridParallelPlugin
      4fa689fc
  23. 08 Dec, 2023 1 commit
  24. 28 Nov, 2023 2 commits
  25. 27 Nov, 2023 1 commit
  26. 22 Nov, 2023 2 commits
  27. 20 Nov, 2023 2 commits
  28. 18 Nov, 2023 1 commit
  29. 16 Nov, 2023 1 commit
    • Elsa Granger's avatar
      [pipeline,shardformer] Fix p2p efficiency in pipeline, allow skipping loading... · b2ad0d9e
      Elsa Granger authored
      
      [pipeline,shardformer] Fix p2p efficiency in pipeline, allow skipping loading weight not in weight_map when `strict=False`, fix llama flash attention forward, add flop estimation by megatron in llama benchmark (#5017)
      
      * Use p2p
      
      * Cannot bidirectonal send p2p
      
      * Refactor tensor creation and serialization in P2P
      communication
      
      * Fix llama forward args in flash attention
      
      * Add flop estimate from megatron
      
      * Support loading weight not in weight_map when strict=False in hybrid_parallel
      
      * Use send_forward_recv_backward, etc in 1f1b
      
      * Use dataclass for metdata
      Remove torch.cuda.synchronize() as suggested
      
      * Add comment about the torch.cuda.synchronize for potential error
      
      * Typo
      
      * Update hybrid_parallel_checkpoint_io.py
      
      * Update p2p.py
      
      * Update one_f_one_b.py
      
      * Update p2p.py
      
      ---------
      Co-authored-by: default avatarflybird11111 <1829166702@qq.com>
      b2ad0d9e
  30. 09 Nov, 2023 1 commit
    • Wenhao Chen's avatar
      [moe]: fix ep/tp tests, add hierarchical all2all (#4982) · 72444127
      Wenhao Chen authored
      * fix: add warning for EP different behavior
      
      * fix: use shard_data in ep & tp model
      
      * to: add used_capacity
      
      * fix: fix router test
      
      * feat: add create_ep_node_group
      
      * feat: add create_ep_hierarchical_group fn
      
      * feat: add HierarchicalAllToAll
      
      * test: add hierarchical all2all test
      
      * fix: fix test errors
      
      * fix: simplify create_ep_hierarchical_group
      
      * fix: add hierarchical_alltoall arg
      
      * fix: fix environ typo
      
      * revert: revert process mesh order
      
      * to: add todo mark
      
      * fix: skip hierarchical_comm if torch < 1.13.1
      72444127
  31. 08 Nov, 2023 1 commit
    • Xuanlei Zhao's avatar
      [moe] support optimizer checkpoint (#5015) · f71e63b0
      Xuanlei Zhao authored
      * Refactor MoE Manager setup method
      
      * unshard optim ckpt
      
      * optim io
      
      * update transformer version
      
      * update requirements
      
      * update ckpt
      
      * update ckpt
      
      * update ckpt
      
      * fix engine
      
      * fix engine
      f71e63b0
  32. 02 Nov, 2023 1 commit
  33. 07 Oct, 2023 1 commit
  34. 21 Sep, 2023 1 commit