1. 25 Mar, 2024 3 commits
    • flybird11111's avatar
      [shardformer]Fix lm parallel. (#5480) · 0688d92e
      flybird11111 authored
      * fix
      
      * padding vocab_size when using pipeline parallellism
      
      padding vocab_size when using pipeline parallellism
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      * fix gather output
      
      * fix
      
      * fix
      
      * fix
      
      fix resize embedding
      
      fix resize embedding
      
      * fix resize embedding
      
      fix
      
      * revert
      
      * revert
      
      * revert
      
      * fix lm forward distribution
      
      * fix
      
      * test ci
      
      * fix
      0688d92e
    • binmakeswell's avatar
      [release] grok-1 inference benchmark (#5500) · 34e90925
      binmakeswell authored
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      
      * [release] grok-1 inference benchmark
      34e90925
    • Wenhao Chen's avatar
      [hotfix] set return_outputs=False in examples and polish code (#5404) · bb0a668f
      Wenhao Chen authored
      * fix: simplify merge_batch
      
      * fix: use return_outputs=False to eliminate extra memory consumption
      
      * feat: add return_outputs warning
      
      * style: remove `return_outputs=False` as it is the default value
      bb0a668f
  2. 24 Mar, 2024 1 commit
    • Yuanheng Zhao's avatar
      [example] update Grok-1 inference (#5495) · 5fcd7795
      Yuanheng Zhao authored
      * revise grok-1 example
      
      * remove unused arg in scripts
      
      * prevent re-installing torch
      
      * update readme
      
      * revert modifying colossalai requirements
      
      * add perf
      
      * trivial
      
      * add tokenizer url
      5fcd7795
  3. 22 Mar, 2024 1 commit
  4. 21 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [example] add grok-1 inference (#5485) · 848a574c
      Hongxin Liu authored
      * [misc] add submodule
      
      * remove submodule
      
      * [example] support grok-1 tp inference
      
      * [example] add grok-1 inference script
      
      * [example] refactor code
      
      * [example] add grok-1 readme
      
      * [exmaple] add test ci
      
      * [exmaple] update readme
      848a574c
  5. 20 Mar, 2024 1 commit
  6. 18 Mar, 2024 2 commits
  7. 13 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [devops] fix compatibility (#5444) · f2e8b9ef
      Hongxin Liu authored
      * [devops] fix compatibility
      
      * [hotfix] update compatibility test on pr
      
      * [devops] fix compatibility
      
      * [devops] record duration during comp test
      
      * [test] decrease test duration
      
      * fix falcon
      f2e8b9ef
  8. 12 Mar, 2024 1 commit
  9. 11 Mar, 2024 1 commit
  10. 07 Mar, 2024 2 commits
  11. 05 Mar, 2024 11 commits
  12. 04 Mar, 2024 1 commit
    • flybird11111's avatar
      [example]add gpt2 benchmark example script. (#5295) · 29695cf7
      flybird11111 authored
      
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed ddp test (#5254)
      
      * [ci] fixed ddp test
      
      * polish
      
      * fix typo in  applications/ColossalEval/README.md (#5250)
      
      * [ci] fix shardformer tests. (#5255)
      
      * fix ci
      
      fix
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [doc] fix doc typo (#5256)
      
      * [doc] fix annotation display
      
      * [doc] fix llama2 doc
      
      * [hotfix]: add pp sanity check and fix mbs arg (#5268)
      
      * fix: fix misleading mbs arg
      
      * feat: add pp sanity check
      
      * fix: fix 1f1b sanity check
      
      * [workflow] fixed incomplete bash command (#5272)
      
      * [workflow] fixed oom tests (#5275)
      
      * [workflow] fixed oom tests
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
      
      * fix ci
      
      fix
      
      * fix test
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      * fix
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
      
      * support gradients acc
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
      
      * fix auto loading gpt2 tokenizer (#5279)
      
      * [doc] add llama2-13B disyplay (#5285)
      
      * Update README.md
      
      * fix 13b typo
      
      ---------
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      
      * fix llama pretrain (#5287)
      
      * fix
      
      * fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      * Update shardformer.py
      
      ---------
      Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
      Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
      Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
      Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
      29695cf7
  13. 01 Mar, 2024 1 commit
  14. 29 Feb, 2024 3 commits
  15. 28 Feb, 2024 1 commit
  16. 27 Feb, 2024 4 commits
  17. 26 Feb, 2024 1 commit
  18. 20 Feb, 2024 1 commit
  19. 19 Feb, 2024 3 commits