• flybird11111's avatar
    [example]add gpt2 benchmark example script. (#5295) · 29695cf7
    flybird11111 authored
    
    
    * benchmark gpt2
    
    * fix
    
    fix
    
    fix
    
    fix
    
    * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
    
    * [workflow] fixed build CI (#5240)
    
    * [workflow] fixed build CI
    
    * polish
    
    * polish
    
    * polish
    
    * polish
    
    * polish
    
    * [ci] fixed booster test (#5251)
    
    * [ci] fixed booster test
    
    * [ci] fixed booster test
    
    * [ci] fixed booster test
    
    * [ci] fixed ddp test (#5254)
    
    * [ci] fixed ddp test
    
    * polish
    
    * fix typo in  applications/ColossalEval/README.md (#5250)
    
    * [ci] fix shardformer tests. (#5255)
    
    * fix ci
    
    fix
    
    * revert: revert p2p
    
    * feat: add enable_metadata_cache option
    
    * revert: enable t5 tests
    
    ---------
    Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
    
    * [doc] fix doc typo (#5256)
    
    * [doc] fix annotation display
    
    * [doc] fix llama2 doc
    
    * [hotfix]: add pp sanity check and fix mbs arg (#5268)
    
    * fix: fix misleading mbs arg
    
    * feat: add pp sanity check
    
    * fix: fix 1f1b sanity check
    
    * [workflow] fixed incomplete bash command (#5272)
    
    * [workflow] fixed oom tests (#5275)
    
    * [workflow] fixed oom tests
    
    * polish
    
    * polish
    
    * polish
    
    * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
    
    * fix ci
    
    fix
    
    * fix test
    
    * revert: revert p2p
    
    * feat: add enable_metadata_cache option
    
    * revert: enable t5 tests
    
    * fix
    
    ---------
    Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
    
    * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
    
    * support gradients acc
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    * fix
    
    fix
    
    * fix
    
    fix
    
    fix
    
    * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
    
    * fix auto loading gpt2 tokenizer (#5279)
    
    * [doc] add llama2-13B disyplay (#5285)
    
    * Update README.md
    
    * fix 13b typo
    
    ---------
    Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
    
    * fix llama pretrain (#5287)
    
    * fix
    
    * fix
    
    * fix
    
    fix
    
    * fix
    
    fix
    
    fix
    
    * fix
    
    fix
    
    * benchmark gpt2
    
    * fix
    
    fix
    
    fix
    
    fix
    
    * [workflow] fixed build CI (#5240)
    
    * [workflow] fixed build CI
    
    * polish
    
    * polish
    
    * polish
    
    * polish
    
    * polish
    
    * [ci] fixed booster test (#5251)
    
    * [ci] fixed booster test
    
    * [ci] fixed booster test
    
    * [ci] fixed booster test
    
    * fix
    
    fix
    
    * fix
    
    fix
    
    fix
    
    * fix
    
    * fix
    
    fix
    
    fix
    
    fix
    
    fix
    
    * fix
    
    * Update shardformer.py
    
    ---------
    Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
    Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
    Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
    Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
    Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
    Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
    Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
    29695cf7
test_gemini_checkpoint_io.py 7.2 KB