"...git@developer.sourcefind.cn:yangql/googletest.git" did not exist on "40cd5d117035c0e280cbeceac42176ffa75a39a3"
- 04 Jul, 2023 14 commits
-
-
wukong1992 authored
test t5
-
wukong1992 authored
adjust layer attr
-
FoolPlayer authored
* add dist dropout in model * update docstring and bert policy with dropout * refactor basepolicy and sharded, update bert * update format * update gpt2 policy * update bert policy * remove unused code * update readme for new policy usage
-
FoolPlayer authored
* fix bug in slicer, add slicer unit test * add dropout test * use pid as dropout seed * updata dropout test with local pattern * ad todo
-
FoolPlayer authored
* add bert align test, fix dist loss bug * forward and backward align * add ignore index * add shardformer CI * add gather_output optional for user in shardconfig * update readme with optional gather_ouput * add dist crossentropy loss test, remove unused files * remove unused file * remove unused file * rename the file * polish code
-
FoolPlayer authored
* add gpt2 policy and modify shard and slicer to support * remove unused code * polish code
-
FoolPlayer authored
-
FoolPlayer authored
* add dropout layer, add dropout test * modify seed manager as context manager * add a copy of col_nn.layer * add dist_crossentropy loss; separate module test * polish the code * fix dist crossentropy loss
-
FoolPlayer authored
* update readme with modules content * remove img
-
Frank Lee authored
* [shardformer] refactored the user api * polish code
-
Frank Lee authored
-
FoolPlayer authored
* init shardformer code structure * add implement of sharder (inject and replace) * add implement of replace layer to colossal layer * separate different layer policy, add some notion * implement 1d and 2d slicer, can tell col or row * fix bug when slicing and inject model * fix some bug; add inference test example * add share weight and train example * add train * add docstring and readme * add docstring for other files * pre-commit
-
FoolPlayer authored
* init shardformer code structure * add implement of sharder (inject and replace) * add implement of replace layer to colossal layer * separate different layer policy, add some notion * implement 1d and 2d slicer, can tell col or row * fix bug when slicing and inject model * fix some bug; add inference test example
-
Wenhao Chen authored
* feat: remove on_learn_epoch fn as not used * revert: add _on_learn_epoch fn * feat: remove NaiveStrategy * test: update train_prompts tests * fix: remove prepare_llama_tokenizer_and_embedding * test: add lora arg * feat: remove roberta support in train_prompts due to runtime errs * feat: remove deberta & roberta in rm as not used * test: remove deberta and roberta tests * feat: remove deberta and roberta models as not used * fix: remove calls to roberta * fix: remove prepare_llama_tokenizer_and_embedding * chore: update transformers version * docs: update transformers version * fix: fix actor inference * fix: fix ci * feat: change llama pad token to unk * revert: revert ddp setup_distributed * fix: change llama pad token to unk * revert: undo unnecessary changes * fix: use pip to install transformers
-
- 03 Jul, 2023 3 commits
-
-
Baizhou Zhang authored
-
digger yu authored
-
digger yu authored
-
- 30 Jun, 2023 1 commit
-
-
digger yu authored
-
- 29 Jun, 2023 2 commits
-
-
Wenhao Chen authored
* feat: remove on_learn_epoch fn as not used * revert: add _on_learn_epoch fn * to: remove the use of NaiveStrategy * test: remove NaiveStrategy tests * feat: remove NaiveStrategy * style: modify comments and params * feat: split ColossalAIStrategy into LowLevelZeroStrategy and GeminiStrategy * fix: remove naive * fix: align with modified colossal strategy * fix: fix ddp _try_init_dist arg
-
Wenhao Chen authored
* to: add SLTrainer * refactor: refactor RMTrainer and SFTTrainer * fix: fix init file * feat: remove on_learn_epoch fn as not used * fix: align with modified gemini arguments * to: add OnPolicyTrainer * revert: add _on_learn_epoch fn * refactor: refactor PPOTrainer * style: rename PPOTrainer argument * fix: align with modified PPO arguments * test: align with modified train_prompts arguments * chore: modify train_prompts * docs: align with modified arguments * fix: remove unnecessary output * fix: move dataloader to fit fn of SLTrainer * fix: move dataloader to fit fn of OnPolicyTrainer * fix: modify usage of prompt and pretrain dataloader
-
- 28 Jun, 2023 4 commits
- 27 Jun, 2023 2 commits
- 26 Jun, 2023 2 commits
-
-
Baizhou Zhang authored
-
Michelle authored
* fix chat eval * fix utils * fix utils * add comment --------- Co-authored-by:Qianran Ma <qianranm@luchentech.com>
-
- 25 Jun, 2023 3 commits
-
-
Baizhou Zhang authored
[gemini] Rename arguments in chunk configuration searching
-
Wenhao Chen authored
* refactor: adapt boost API in base and naive strategies * fix: initialize plugin after setup_distributed * fix: fix save_pretrained fn * refactor: adapt boost API in DDPStrategy * to: add _post_init check * to: fix ddp backward, modify ddp dataloader and unwrap * feat: adapt boost API in ColossalAIStrategy * fix: call setup_distributed before use get_current_device * fix: fix save_model and save_optimizer * test: remove save_sharded_optimizer test * style: apply formatter * fix: fix stage check and add comments * feat: allow dict type arg in strategy.prepare * to: temporarily remove lr_scheduler for testing * style: simplify init of ColossalAIStrategy * fix: fix lr_scheduler in sft and rm * style: modify comments * test: add train_prompts tests * fix: fix inference only case and use in train_prompts * test: skip failed tests in ci * style: fix CodeFactor check * fix: do not use model.to('cpu') with GeminiPlugin * test: enable colossalai_gemini tests * test: set CUDA_VISIBLE_DEVICES in ci * docs: add note -
Baizhou Zhang authored
-
- 22 Jun, 2023 1 commit
-
-
Frank Lee authored
-
- 19 Jun, 2023 5 commits
-
-
Hongxin Liu authored
* [devops] fix build on pr ci * [devops] fix build on pr ci
-
github-actions[bot] authored
Co-authored-by:github-actions <github-actions@github.com>
-
LuGY authored
-
digger yu authored
-
Frank Lee authored
[sync] sync develop to main
-
- 16 Jun, 2023 2 commits
-
-
Frank Lee authored
-
Baizhou Zhang authored
-
- 15 Jun, 2023 1 commit
-
-
Wenhao Chen authored
* feat: make optimizer optional in Booster.boost * test: skip unet test if diffusers version > 0.10.2
-