- 05 Mar, 2024 2 commits
-
-
Hongxin Liu authored
-
binmakeswell authored
* [doc] sora release * [doc] sora release * [doc] sora release * [doc] sora release
-
- 29 Feb, 2024 1 commit
-
-
binmakeswell authored
-
- 19 Feb, 2024 2 commits
- 25 Jan, 2024 1 commit
-
-
digger yu authored
-
- 09 Jan, 2024 1 commit
-
-
Hongxin Liu authored
* update accelerator * fix timer * fix amp * update * fix * update bug * add error raise * fix autocast * fix set device * remove doc accelerator * update doc * update doc * update doc * use nullcontext * update cpu * update null context * change time limit for example * udpate * update * update * update * [npu] polish accelerator code --------- Co-authored-by:
Xuanlei Zhao <xuanlei.zhao@gmail.com> Co-authored-by:
zxl <43881818+oahzxl@users.noreply.github.com>
-
- 08 Jan, 2024 1 commit
-
-
binmakeswell authored
* [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release
-
- 07 Jan, 2024 1 commit
-
-
binmakeswell authored
* [doc] add Colossal-LLaMA-2-13B * [doc] add Colossal-LLaMA-2-13B * [doc] add Colossal-LLaMA-2-13B
-
- 15 Dec, 2023 1 commit
-
-
flybird11111 authored
* fix aaa fix fix fix * fix * fix * test ci * fix ci fix * update pytorch version in documents
-
- 28 Nov, 2023 2 commits
-
-
binmakeswell authored
* [doc] add moe news * [doc] add moe news * [doc] add moe news
-
Wenhao Chen authored
* [shardformer] implement policy for all GPT-J models and test * [shardformer] support interleaved pipeline parallel for bert finetune * [shardformer] shardformer support falcon (#4883) * [shardformer]: fix interleaved pipeline for bert model (#5048) * [hotfix]: disable seq parallel for gptj and falcon, and polish code (#5093) * Add Mistral support for Shardformer (#5103) * [shardformer] add tests to mistral (#5105) --------- Co-authored-by:
Pengtai Xu <henryxu880@gmail.com> Co-authored-by:
ppt0011 <143150326+ppt0011@users.noreply.github.com> Co-authored-by:
flybird11111 <1829166702@qq.com> Co-authored-by:
eric8607242 <e0928021388@gmail.com>
-
- 27 Nov, 2023 1 commit
-
-
digger yu authored
-
- 24 Nov, 2023 1 commit
-
-
digger yu authored
-
- 22 Nov, 2023 1 commit
-
-
digger yu authored
-
- 21 Nov, 2023 1 commit
-
-
digger yu authored
-
- 31 Oct, 2023 1 commit
-
-
ppt0011 authored
-
- 18 Oct, 2023 1 commit
-
-
digger yu authored
-
- 17 Oct, 2023 1 commit
-
-
Baizhou Zhang authored
* add test * fix no_sync bug in low level zero plugin * fix test * add argument for grad accum * add grad accum in backward hook for gemini * finish implementation, rewrite tests * fix test * skip stuck model in low level zero test * update doc * optimize communication & fix gradient checkpoint * modify doc * cleaning codes * update cpu adam fp16 case
-
- 10 Oct, 2023 1 commit
-
-
flybird11111 authored
* [doc]update advanced tutorials, training gpt with hybrid parallelism * [doc]update advanced tutorials, training gpt with hybrid parallelism * update vit tutorials * update vit tutorials * update vit tutorials * update vit tutorials * update en/train_vit_with_hybrid_parallel.py * fix * resolve comments * fix
-
- 05 Oct, 2023 1 commit
-
-
Zhongkai Zhao authored
-
- 27 Sep, 2023 2 commits
-
-
binmakeswell authored
-
Hongxin Liu authored
-
- 26 Sep, 2023 2 commits
-
-
Baizhou Zhang authored
* support unsharded saving/loading for model * support optimizer unsharded saving * update doc * support unsharded loading for optimizer * small fix
-
Baizhou Zhang authored
* fix example format in docstring * polish shardformer doc
-
- 25 Sep, 2023 1 commit
-
-
binmakeswell authored
* [doc] add llama2 domain-specific solution news
-
- 21 Sep, 2023 2 commits
-
-
Baizhou Zhang authored
-
Hongxin Liu authored
* [doc] clean up outdated docs * [doc] fix linking * [doc] fix linking
-
- 20 Sep, 2023 1 commit
-
-
Pengtai Xu authored
-
- 19 Sep, 2023 3 commits
-
-
Pengtai Xu authored
-
Pengtai Xu authored
-
Pengtai Xu authored
-
- 18 Sep, 2023 1 commit
-
-
Hongxin Liu authored
* [legacy] remove outdated codes of pipeline (#4692) * [legacy] remove cli of benchmark and update optim (#4690) * [legacy] remove cli of benchmark and update optim * [doc] fix cli doc test * [legacy] fix engine clip grad norm * [legacy] remove outdated colo tensor (#4694) * [legacy] remove outdated colo tensor * [test] fix test import * [legacy] move outdated zero to legacy (#4696) * [legacy] clean up utils (#4700) * [legacy] clean up utils * [example] update examples * [legacy] clean up amp * [legacy] fix amp module * [legacy] clean up gpc (#4742) * [legacy] clean up context * [legacy] clean core, constants and global vars * [legacy] refactor initialize * [example] fix examples ci * [example] fix examples ci * [legacy] fix tests * [example] fix gpt example * [example] fix examples ci * [devops] fix ci installation * [example] fix examples ci
-
- 15 Sep, 2023 7 commits
-
-
Baizhou Zhang authored
-
Baizhou Zhang authored
* arrange position of chapters * fix typos in seq parallel doc
-
Bin Jia authored
* update doc of seq parallel * fix typo
-
flybird11111 authored
* [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document * [shardformer] update pipeline parallel document
-
Baizhou Zhang authored
* add compatibility matrix for shardformer doc * update tp doc
-
github-actions[bot] authored
Co-authored-by:github-actions <github-actions@github.com>
-
Baizhou Zhang authored
* create shardformer doc files * add docstring for seq-parallel * update ShardConfig docstring * add links to llama example * add outdated massage * finish introduction & supporting information * finish 'how shardformer works' * finish shardformer.md English doc * fix doctest fail * add Chinese document
-