- 05 Mar, 2024 3 commits
-
-
MickeyCHAN authored
* fix import error * Update dpt_depth.py --------- Co-authored-by:binmakeswell <binmakeswell@gmail.com>
-
Hongxin Liu authored
-
binmakeswell authored
* [doc] sora release * [doc] sora release * [doc] sora release * [doc] sora release
-
- 04 Mar, 2024 1 commit
-
-
flybird11111 authored
* benchmark gpt2 * fix fix fix fix * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247) * [workflow] fixed build CI (#5240) * [workflow] fixed build CI * polish * polish * polish * polish * polish * [ci] fixed booster test (#5251) * [ci] fixed booster test * [ci] fixed booster test * [ci] fixed booster test * [ci] fixed ddp test (#5254) * [ci] fixed ddp test * polish * fix typo in applications/ColossalEval/README.md (#5250) * [ci] fix shardformer tests. (#5255) * fix ci fix * revert: revert p2p * feat: add enable_metadata_cache option * revert: enable t5 tests --------- Co-authored-by:
Wenhao Chen <cwher@outlook.com> * [doc] fix doc typo (#5256) * [doc] fix annotation display * [doc] fix llama2 doc * [hotfix]: add pp sanity check and fix mbs arg (#5268) * fix: fix misleading mbs arg * feat: add pp sanity check * fix: fix 1f1b sanity check * [workflow] fixed incomplete bash command (#5272) * [workflow] fixed oom tests (#5275) * [workflow] fixed oom tests * polish * polish * polish * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276) * fix ci fix * fix test * revert: revert p2p * feat: add enable_metadata_cache option * revert: enable t5 tests * fix --------- Co-authored-by:
Wenhao Chen <cwher@outlook.com> * [shardformer] hybridparallelplugin support gradients accumulation. (#5246) * support gradients acc fix fix fix fix fix fix fix fix fix fix fix fix fix * fix fix * fix fix fix * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230) * fix auto loading gpt2 tokenizer (#5279) * [doc] add llama2-13B disyplay (#5285) * Update README.md * fix 13b typo --------- Co-authored-by:
binmakeswell <binmakeswell@gmail.com> * fix llama pretrain (#5287) * fix * fix * fix fix * fix fix fix * fix fix * benchmark gpt2 * fix fix fix fix * [workflow] fixed build CI (#5240) * [workflow] fixed build CI * polish * polish * polish * polish * polish * [ci] fixed booster test (#5251) * [ci] fixed booster test * [ci] fixed booster test * [ci] fixed booster test * fix fix * fix fix fix * fix * fix fix fix fix fix * fix * Update shardformer.py --------- Co-authored-by:
digger yu <digger-yu@outlook.com> Co-authored-by:
Frank Lee <somerlee.9@gmail.com> Co-authored-by:
Wenhao Chen <cwher@outlook.com> Co-authored-by:
binmakeswell <binmakeswell@gmail.com> Co-authored-by:
Zhongkai Zhao <kanezz620@gmail.com> Co-authored-by:
Michelle <97082656+MichelleMa8@users.noreply.github.com> Co-authored-by:
Desperado-Jia <502205863@qq.com>
-
- 01 Mar, 2024 1 commit
-
-
Camille Zhong authored
-
- 29 Feb, 2024 3 commits
-
-
binmakeswell authored
-
binmakeswell authored
-
Frank Lee authored
-
- 28 Feb, 2024 1 commit
-
-
Tong Li authored
-
- 27 Feb, 2024 4 commits
-
-
flybird11111 authored
* gather llama logits * fix
-
Frank Lee authored
-
QinLuo authored
-
Hongxin Liu authored
-
- 26 Feb, 2024 1 commit
-
-
Hongxin Liu authored
-
- 20 Feb, 2024 1 commit
-
-
Stephan Kölker authored
-
- 19 Feb, 2024 4 commits
-
-
CZYCW authored
Co-authored-by:binmakeswell <binmakeswell@gmail.com>
-
Frank Lee authored
-
yixiaoer authored
-
Hongxin Liu authored
* [llama] refactor inference example to fit sft * [llama] fix training script to fit gemini * [llama] fix inference script
-
- 08 Feb, 2024 4 commits
-
-
Hongxin Liu authored
-
Frank Lee authored
-
Frank Lee authored
[llama] support npu for Colossal-LLaMA-2
-
ver217 authored
-
- 07 Feb, 2024 6 commits
-
-
Hongxin Liu authored
-
Hongxin Liu authored
-
Hongxin Liu authored
-
Hongxin Liu authored
* [moe] add mixtral block for single expert * [moe] mixtral block fwd support uneven ep * [moe] mixtral block bwd support uneven ep * [moe] add mixtral moe layer * [moe] simplify replace * [meo] support save sharded mixtral * [meo] support load sharded mixtral * [meo] support save sharded optim * [meo] integrate moe manager into plug * [meo] fix optimizer load * [meo] fix mixtral layer
-
Hongxin Liu authored
* [moe] top2 allow uneven input * [moe] update capacity computing * [moe] remove debug info * [moe] update capacity computing * [moe] update capacity computing
-
Xuanlei Zhao authored
-
- 06 Feb, 2024 4 commits
-
-
Hongxin Liu authored
* [llama] fix memory issue * [llama] add comment
-
Hongxin Liu authored
-
Hongxin Liu authored
-
Camille Zhong authored
-
- 05 Feb, 2024 4 commits
-
-
Camille Zhong authored
-
Hongxin Liu authored
-
Hongxin Liu authored
* [llama] update training script * [doc] polish docstr
-
Hongxin Liu authored
* [plugin] refactor prepare dataloader * [plugin] update train script
-
- 04 Feb, 2024 1 commit
-
-
Hongxin Liu authored
* [gemini] fix param op hook when output is tuple * [gemini] fix param op hook
-
- 02 Feb, 2024 1 commit
-
-
Wenhao Chen authored
* fix: remove unnecessary assert * test: add more 3d plugin tests * fix: add warning
-
- 01 Feb, 2024 1 commit
-
-
Hongxin Liu authored
* [checkpointio] fix hybrid parallel optim checkpoint * [extension] fix cuda extension * [checkpointio] fix gemini optimizer checkpoint * polish code
-