1. 02 Apr, 2024 1 commit
  2. 01 Apr, 2024 1 commit
    • Wenhao Chen's avatar
      [shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous... · e614aa34
      Wenhao Chen authored
      [shardformer, pipeline] add `gradient_checkpointing_ratio` and heterogenous shard policy for llama (#5508)
      
      * feat: add `GradientCheckpointConfig` and `PipelineGradientCheckpointConfig`
      
      * feat: apply `GradientCheckpointConfig` to policy and llama_forward
      
      * feat: move `distribute_layer` and `get_stage_index` to PipelineStageManager
      
      * fix: add optional args for `distribute_layer` and `get_stage_index`
      
      * fix: fix changed API calls
      
      * test: update llama tests
      
      * style: polish `GradientCheckpointConfig`
      
      * fix: fix pipeline utils tests
      e614aa34
  3. 29 Mar, 2024 1 commit
    • YeAnbang's avatar
      [ColossalChat] Update RLHF V2 (#5286) · df5e9c53
      YeAnbang authored
      
      
      * Add dpo. Fix sft, ppo, lora. Refactor all
      
      * fix and tested ppo
      
      * 2 nd round refactor
      
      * add ci tests
      
      * fix ci
      
      * fix ci
      
      * fix readme, style
      
      * fix readme style
      
      * fix style, fix benchmark
      
      * reproduce benchmark result, remove useless files
      
      * rename to ColossalChat
      
      * use new image
      
      * fix ci workflow
      
      * fix ci
      
      * use local model/tokenizer for ci tests
      
      * fix ci
      
      * fix ci
      
      * fix ci
      
      * fix ci timeout
      
      * fix rm progress bar. fix ci timeout
      
      * fix ci
      
      * fix ci typo
      
      * remove 3d plugin from ci temporary
      
      * test environment
      
      * cannot save optimizer
      
      * support chat template
      
      * fix readme
      
      * fix path
      
      * test ci locally
      
      * restore build_or_pr
      
      * fix ci data path
      
      * fix benchmark
      
      * fix ci, move ci tests to 3080, disable fast tokenizer
      
      * move ci to 85
      
      * support flash attention 2
      
      * add all-in-one data preparation script. Fix colossal-llama2-chat chat template
      
      * add hardware requirements
      
      * move ci test data
      
      * fix save_model, add unwrap
      
      * fix missing bos
      
      * fix missing bos; support grad accumulation with gemini
      
      * fix ci
      
      * fix ci
      
      * fix ci
      
      * fix llama2 chat template config
      
      * debug sft
      
      * debug sft
      
      * fix colossalai version requirement
      
      * fix ci
      
      * add sanity check to prevent NaN loss
      
      * fix requirements
      
      * add dummy data generation script
      
      * add dummy data generation script
      
      * add dummy data generation script
      
      * add dummy data generation script
      
      * update readme
      
      * update readme
      
      * update readme and ignore
      
      * fix logger bug
      
      * support parallel_output
      
      * modify data preparation logic
      
      * fix tokenization
      
      * update lr
      
      * fix inference
      
      * run pre-commit
      
      ---------
      Co-authored-by: default avatarTong Li <tong.li352711588@gmail.com>
      df5e9c53
  4. 28 Mar, 2024 1 commit
  5. 27 Mar, 2024 3 commits
    • Insu Jang's avatar
      [shardformer] fix pipeline forward error if custom layer distribution is used (#5189) · 00525f77
      Insu Jang authored
      
      
      * Use self.[distribute_layers|get_stage_index] to exploit custom layer distribution
      
      * Change static methods for t5 layer distribution to member functions
      
      * Change static methods for whisper layer distribution to member functions
      
      * Replace whisper policy usage with self one
      
      * Fix test case to use non-static layer distribution methods
      
      * fix: fix typo
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      00525f77
    • github-actions[bot]'s avatar
    • Hongxin Liu's avatar
      [shardformer] update colo attention to support custom mask (#5510) · 19e1a5cf
      Hongxin Liu authored
      * [feature] refactor colo attention (#5462)
      
      * [extension] update api
      
      * [feature] add colo attention
      
      * [feature] update sdpa
      
      * [feature] update npu attention
      
      * [feature] update flash-attn
      
      * [test] add flash attn test
      
      * [test] update flash attn test
      
      * [shardformer] update modeling to fit colo attention (#5465)
      
      * [misc] refactor folder structure
      
      * [shardformer] update llama flash-attn
      
      * [shardformer] fix llama policy
      
      * [devops] update tensornvme install
      
      * [test] update llama test
      
      * [shardformer] update colo attn kernel dispatch
      
      * [shardformer] update blip2
      
      * [shardformer] update chatglm
      
      * [shardformer] update gpt2
      
      * [shardformer] update gptj
      
      * [shardformer] update opt
      
      * [shardformer] update vit
      
      * [shardformer] update colo attention mask prep
      
      * [shardformer] update whisper
      
      * [test] fix shardformer tests (#5514)
      
      * [test] fix shardformer tests
      
      * [test] fix shardformer tests
      19e1a5cf
  6. 26 Mar, 2024 6 commits
  7. 25 Mar, 2024 3 commits
    • flybird11111's avatar
      [shardformer]Fix lm parallel. (#5480) · 0688d92e
      flybird11111 authored
      * fix
      
      * padding vocab_size when using pipeline parallellism
      
      padding vocab_size when using pipeline parallellism
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      * fix gather output
      
      * fix
      
      * fix
      
      * fix
      
      fix resize embedding
      
      fix resize embedding
      
      * fix resize embedding
      
      fix
      
      * revert
      
      * revert
      
      * revert
      
      * fix lm forward distribution
      
      * fix
      
      * test ci
      
      * fix
      0688d92e
    • binmakeswell's avatar
      [release] grok-1 inference benchmark (#5500) · 34e90925
      binmakeswell authored
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      34e90925
    • Wenhao Chen's avatar
      [hotfix] set return_outputs=False in examples and polish code (#5404) · bb0a668f
      Wenhao Chen authored
      * fix: simplify merge_batch
      
      * fix: use return_outputs=False to eliminate extra memory consumption
      
      * feat: add return_outputs warning
      
      * style: remove `return_outputs=False` as it is the default value
      bb0a668f
  8. 24 Mar, 2024 1 commit
    • Yuanheng Zhao's avatar
      [example] update Grok-1 inference (#5495) · 5fcd7795
      Yuanheng Zhao authored
      * revise grok-1 example
      
      * remove unused arg in scripts
      
      * prevent re-installing torch
      
      * update readme
      
      * revert modifying colossalai requirements
      
      * add perf
      
      * trivial
      
      * add tokenizer url
      5fcd7795
  9. 22 Mar, 2024 1 commit
  10. 21 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [example] add grok-1 inference (#5485) · 848a574c
      Hongxin Liu authored
      * [misc] add submodule
      
      * remove submodule
      
      * [example] support grok-1 tp inference
      
      * [example] add grok-1 inference script
      
      * [example] refactor code
      
      * [example] add grok-1 readme
      
      * [exmaple] add test ci
      
      * [exmaple] update readme
      848a574c
  11. 20 Mar, 2024 1 commit
  12. 18 Mar, 2024 2 commits
  13. 13 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [devops] fix compatibility (#5444) · f2e8b9ef
      Hongxin Liu authored
      * [devops] fix compatibility
      
      * [hotfix] update compatibility test on pr
      
      * [devops] fix compatibility
      
      * [devops] record duration during comp test
      
      * [test] decrease test duration
      
      * fix falcon
      f2e8b9ef
  14. 12 Mar, 2024 1 commit
  15. 11 Mar, 2024 1 commit
  16. 07 Mar, 2024 2 commits
  17. 05 Mar, 2024 11 commits
  18. 04 Mar, 2024 1 commit
    • flybird11111's avatar
      [example]add gpt2 benchmark example script. (#5295) · 29695cf7
      flybird11111 authored
      
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed ddp test (#5254)
      
      * [ci] fixed ddp test
      
      * polish
      
      * fix typo in  applications/ColossalEval/README.md (#5250)
      
      * [ci] fix shardformer tests. (#5255)
      
      * fix ci
      
      fix
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [doc] fix doc typo (#5256)
      
      * [doc] fix annotation display
      
      * [doc] fix llama2 doc
      
      * [hotfix]: add pp sanity check and fix mbs arg (#5268)
      
      * fix: fix misleading mbs arg
      
      * feat: add pp sanity check
      
      * fix: fix 1f1b sanity check
      
      * [workflow] fixed incomplete bash command (#5272)
      
      * [workflow] fixed oom tests (#5275)
      
      * [workflow] fixed oom tests
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
      
      * fix ci
      
      fix
      
      * fix test
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      * fix
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
      
      * support gradients acc
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
      
      * fix auto loading gpt2 tokenizer (#5279)
      
      * [doc] add llama2-13B disyplay (#5285)
      
      * Update README.md
      
      * fix 13b typo
      
      ---------
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      
      * fix llama pretrain (#5287)
      
      * fix
      
      * fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      * Update shardformer.py
      
      ---------
      Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
      Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
      Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
      Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
      29695cf7
  19. 01 Mar, 2024 1 commit