- 09 Jun, 2022 1 commit
-
-
Frank Lee authored
* [test] skip tests when not enough GPUs are detected * polish code * polish code
-
- 08 Jun, 2022 1 commit
-
-
Frank Lee authored
* [test] ignore 8 gpu test * polish code * polish workflow * polish workflow
-
- 06 Jun, 2022 1 commit
-
-
Jiarui Fang authored
-
- 24 Apr, 2022 2 commits
-
-
HELSON authored
* refactor StatefulTensor, tensor utilities * add unitest for GeminiMemoryManager
-
YuliangLiu0306 authored
* [CLI] add CLI launcher * Revert "[CLI] add CLI launcher" This reverts commit df7e6506d4500af6a9220ef7fe4d3c7b1daebd4c. * [pipeline]add module lazy init feature to support large model initization. * [pipeline]add to_layer_list and partition method to support arbitrary non-pp model * refactor the module structure * polish * [pipelinable]add unit test for pipelinable * polish * polish * Fix CodeFactor issues.
-
- 19 Apr, 2022 1 commit
-
-
Jiarui Fang authored
-
- 14 Apr, 2022 1 commit
-
-
Frank Lee authored
* [test] refactored with the new rerun decorator * polish test case
-
- 12 Apr, 2022 2 commits
-
-
Jiarui Fang authored
-
FrankLeeeee authored
-
- 11 Apr, 2022 2 commits
-
-
Jiarui Fang authored
-
Jiarui Fang authored
-
- 02 Apr, 2022 1 commit
-
-
HELSON authored
* remove hybrid adam in test_moe_zero_optim * fix activation checkpointing and its unitest
-
- 01 Apr, 2022 3 commits
-
-
アマデウス authored
-
FredHuang99 authored
-
Jiarui Fang authored
-
- 28 Mar, 2022 1 commit
-
-
Jiarui Fang authored
-
- 25 Mar, 2022 5 commits
-
-
Jiarui Fang authored
-
Frank Lee authored
-
Jiarui Fang authored
-
Jiarui Fang authored
-
Jiarui Fang authored
-
- 24 Mar, 2022 2 commits
-
-
Jiarui Fang authored
-
Jiarui Fang authored
-
- 22 Mar, 2022 1 commit
-
-
Jiarui Fang authored
* [zero] polish sharded param name * polish code * polish * polish code * polish * polsih * polish
-
- 18 Mar, 2022 1 commit
-
-
Frank Lee authored
-
- 14 Mar, 2022 3 commits
-
-
Jiarui Fang authored
-
Jiarui Fang authored
-
LuGY authored
* Added tensor detector * Added the - states * Allowed change include_cpu when detect()
-
- 11 Mar, 2022 7 commits
-
-
Frank Lee authored
-
Frank Lee authored
* refactored test with component func * fixed bug
-
LuGY authored
* Added activation offload * Fixed the import bug, used the pytest
-
Jiarui Fang authored
-
jiaruifang authored
This reverts commit bef05489b642385c80e59fe757d598efd1752ecf.
-
Jiarui Fang authored
-
Jiarui Fang authored
* add zero1 (#209) * add zero1 * add test zero1 * update zero stage 1 develop (#212) * Implement naive zero3 (#240) * naive zero3 works well * add zero3 param manager * add TODOs in comments * add gather full param ctx * fix sub module streams * add offload * fix bugs of hook and add unit tests * fix bugs of hook and add unit tests (#252) * add gather full param ctx * fix sub module streams * add offload * fix bugs of hook and add unit tests * polish code and add state dict hook * fix bug * update unit test * refactor reconstructed zero code * clip_grad support zero3 and add unit test * add unit test for Zero3ParameterManager * [WIP] initialize the shard param class * [WIP] Yet another sharded model implementation (#274) * [WIP] initialize the shard param class * [WIP] Yes another implementation of shardModel. Using a better hook method. * torch.concat -> torch.cat * fix test_zero_level_1.py::test_zero_level_1 unitest * remove deepspeed implementation and refactor for the reconstructed zero module * polish zero dp unittests Co-authored-by:
ver217 <lhx0217@gmail.com> Co-authored-by:
Frank Lee <somerlee.9@gmail.com>
-
- 29 Dec, 2021 1 commit
-
-
アマデウス authored
* optimized 1d layer apis; reorganized nn.layer modules; fixed tests * fixed 2.5d runtime issue * reworked split batch, now called in trainer.schedule.load_batch Co-authored-by:BoxiangW <45734921+BoxiangW@users.noreply.github.com>
-
- 16 Dec, 2021 1 commit
-
-
Frank Lee authored
-
- 09 Dec, 2021 1 commit
-
-
Frank Lee authored
* Add gradient accumulation, fix lr scheduler * fix FP16 optimizer and adapted torch amp with tensor parallel (#18) * fixed bugs in compatibility between torch amp and tensor parallel and performed some minor fixes * fixed trainer * Revert "fixed trainer" This reverts commit 2e0b0b76990e8d4e337add483d878c0f61cf5097. * improved consistency between trainer, engine and schedule (#23) Co-authored-by:
1SAA <c2h214748@gmail.com> * Split conv2d, class token, positional embedding in 2d, Fix random number in ddp Fix convergence in cifar10, Imagenet1000 * Integrate 1d tensor parallel in Colossal-AI (#39) * fixed 1D and 2D convergence (#38) * optimized 2D operations * fixed 1D ViT convergence problem * Feature/ddp (#49) * remove redundancy func in setup (#19) (#20) * use env to control the language of doc (#24) (#25) * Support TP-compatible Torch AMP and Update trainer API (#27) * Add gradient accumulation, fix lr scheduler * fix FP16 optimizer and adapted torch amp with tensor parallel (#18) * fixed bugs in compatibility between torch amp and tensor parallel and performed some minor fixes * fixed trainer * Revert "fixed trainer" This reverts commit 2e0b0b76990e8d4e337add483d878c0f61cf5097. * improved consistency between trainer, engine and schedule (#23) Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
ver217 <lhx0217@gmail.com> * add an example of ViT-B/16 and remove w_norm clipping in LAMB (#29) * add explanation for ViT example (#35) (#36) * support torch ddp * fix loss accumulation * add log for ddp * change seed * modify timing hook Co-authored-by:
Frank Lee <somerlee.9@gmail.com> Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
binmakeswell <binmakeswell@gmail.com> * Feature/pipeline (#40) * remove redundancy func in setup (#19) (#20) * use env to control the language of doc (#24) (#25) * Support TP-compatible Torch AMP and Update trainer API (#27) * Add gradient accumulation, fix lr scheduler * fix FP16 optimizer and adapted torch amp with tensor parallel (#18) * fixed bugs in compatibility between torch amp and tensor parallel and performed some minor fixes * fixed trainer * Revert "fixed trainer" This reverts commit 2e0b0b76990e8d4e337add483d878c0f61cf5097. * improved consistency between trainer, engine and schedule (#23) Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
ver217 <lhx0217@gmail.com> * add an example of ViT-B/16 and remove w_norm clipping in LAMB (#29) * add explanation for ViT example (#35) (#36) * optimize communication of pipeline parallel * fix grad clip for pipeline Co-authored-by:
Frank Lee <somerlee.9@gmail.com> Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
binmakeswell <binmakeswell@gmail.com> * optimized 3d layer to fix slow computation ; tested imagenet performance with 3d; reworked lr_scheduler config definition; fixed launch args; fixed some printing issues; simplified apis of 3d layers (#51) * Update 2.5d layer code to get a similar accuracy on imagenet-1k dataset * update api for better usability (#58) update api for better usability Co-authored-by:
1SAA <c2h214748@gmail.com> Co-authored-by:
ver217 <lhx0217@gmail.com> Co-authored-by:
puck_WCR <46049915+WANG-CR@users.noreply.github.com> Co-authored-by:
binmakeswell <binmakeswell@gmail.com> Co-authored-by:
アマデウス <kurisusnowdeng@users.noreply.github.com> Co-authored-by:
BoxiangW <45734921+BoxiangW@users.noreply.github.com>
-
- 28 Oct, 2021 1 commit
-
-
zbian authored
-