- 26 Apr, 2024 1 commit
-
-
binmakeswell authored
* [news] llama3 and open-sora v1.1 * [news] llama3 and open-sora v1.1
-
- 25 Apr, 2024 1 commit
-
-
Hongxin Liu authored
* [shardformer] fix chatglm policy * [shardformer] fix chatglm flash attn * [shardformer] update readme * [shardformer] fix chatglm init * [shardformer] fix chatglm test * [pipeline] fix chatglm merge batch
-
- 23 Apr, 2024 1 commit
-
-
binmakeswell authored
* release llama3 * [release] llama3 * [release] llama3 * [release] llama3 * [release] llama3
-
- 08 Apr, 2024 1 commit
-
-
Hongxin Liu authored
* [devops] remove post commit ci * [misc] run pre-commit on all files * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --------- Co-authored-by:
pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
-
- 25 Mar, 2024 2 commits
-
-
binmakeswell authored
* [release] grok-1 inference benchmark * [release] grok-1 inference benchmark * [release] grok-1 inference benchmark * [release] grok-1 inference benchmark * [release] grok-1 inference benchmark
-
Wenhao Chen authored
* fix: simplify merge_batch * fix: use return_outputs=False to eliminate extra memory consumption * feat: add return_outputs warning * style: remove `return_outputs=False` as it is the default value
-
- 22 Mar, 2024 1 commit
-
-
binmakeswell authored
* [release] grok-1 inference * [release] grok-1 inference * [release] grok-1 inference
-
- 20 Mar, 2024 1 commit
-
-
binmakeswell authored
* [doc] update open-sora demo * [doc] update open-sora demo * [doc] update open-sora demo
-
- 18 Mar, 2024 1 commit
-
-
binmakeswell authored
* [doc] release Open-Sora 1.0 with model weights * [doc] release Open-Sora 1.0 with model weights * [doc] release Open-Sora 1.0 with model weights
-
- 05 Mar, 2024 3 commits
-
-
digger yu authored
-
Hongxin Liu authored
-
binmakeswell authored
* [doc] sora release * [doc] sora release * [doc] sora release * [doc] sora release
-
- 29 Feb, 2024 1 commit
-
-
binmakeswell authored
-
- 19 Feb, 2024 2 commits
- 25 Jan, 2024 1 commit
-
-
digger yu authored
-
- 09 Jan, 2024 1 commit
-
-
Hongxin Liu authored
* update accelerator * fix timer * fix amp * update * fix * update bug * add error raise * fix autocast * fix set device * remove doc accelerator * update doc * update doc * update doc * use nullcontext * update cpu * update null context * change time limit for example * udpate * update * update * update * [npu] polish accelerator code --------- Co-authored-by:
Xuanlei Zhao <xuanlei.zhao@gmail.com> Co-authored-by:
zxl <43881818+oahzxl@users.noreply.github.com>
-
- 08 Jan, 2024 1 commit
-
-
binmakeswell authored
* [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release * [doc] SwiftInfer release
-
- 07 Jan, 2024 1 commit
-
-
binmakeswell authored
* [doc] add Colossal-LLaMA-2-13B * [doc] add Colossal-LLaMA-2-13B * [doc] add Colossal-LLaMA-2-13B
-
- 15 Dec, 2023 1 commit
-
-
flybird11111 authored
* fix aaa fix fix fix * fix * fix * test ci * fix ci fix * update pytorch version in documents
-
- 28 Nov, 2023 2 commits
-
-
binmakeswell authored
* [doc] add moe news * [doc] add moe news * [doc] add moe news
-
Wenhao Chen authored
* [shardformer] implement policy for all GPT-J models and test * [shardformer] support interleaved pipeline parallel for bert finetune * [shardformer] shardformer support falcon (#4883) * [shardformer]: fix interleaved pipeline for bert model (#5048) * [hotfix]: disable seq parallel for gptj and falcon, and polish code (#5093) * Add Mistral support for Shardformer (#5103) * [shardformer] add tests to mistral (#5105) --------- Co-authored-by:
Pengtai Xu <henryxu880@gmail.com> Co-authored-by:
ppt0011 <143150326+ppt0011@users.noreply.github.com> Co-authored-by:
flybird11111 <1829166702@qq.com> Co-authored-by:
eric8607242 <e0928021388@gmail.com>
-
- 27 Nov, 2023 1 commit
-
-
digger yu authored
-
- 24 Nov, 2023 1 commit
-
-
digger yu authored
-
- 22 Nov, 2023 1 commit
-
-
digger yu authored
-
- 21 Nov, 2023 1 commit
-
-
digger yu authored
-
- 31 Oct, 2023 1 commit
-
-
ppt0011 authored
-
- 18 Oct, 2023 1 commit
-
-
digger yu authored
-
- 17 Oct, 2023 1 commit
-
-
Baizhou Zhang authored
* add test * fix no_sync bug in low level zero plugin * fix test * add argument for grad accum * add grad accum in backward hook for gemini * finish implementation, rewrite tests * fix test * skip stuck model in low level zero test * update doc * optimize communication & fix gradient checkpoint * modify doc * cleaning codes * update cpu adam fp16 case
-
- 10 Oct, 2023 1 commit
-
-
flybird11111 authored
* [doc]update advanced tutorials, training gpt with hybrid parallelism * [doc]update advanced tutorials, training gpt with hybrid parallelism * update vit tutorials * update vit tutorials * update vit tutorials * update vit tutorials * update en/train_vit_with_hybrid_parallel.py * fix * resolve comments * fix
-
- 05 Oct, 2023 1 commit
-
-
Zhongkai Zhao authored
-
- 27 Sep, 2023 2 commits
-
-
binmakeswell authored
-
Hongxin Liu authored
-
- 26 Sep, 2023 2 commits
-
-
Baizhou Zhang authored
* support unsharded saving/loading for model * support optimizer unsharded saving * update doc * support unsharded loading for optimizer * small fix
-
Baizhou Zhang authored
* fix example format in docstring * polish shardformer doc
-
- 25 Sep, 2023 1 commit
-
-
binmakeswell authored
* [doc] add llama2 domain-specific solution news
-
- 21 Sep, 2023 2 commits
-
-
Baizhou Zhang authored
-
Hongxin Liu authored
* [doc] clean up outdated docs * [doc] fix linking * [doc] fix linking
-
- 20 Sep, 2023 1 commit
-
-
Pengtai Xu authored
-
- 19 Sep, 2023 1 commit
-
-
Pengtai Xu authored
-