1. 27 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [shardformer] update colo attention to support custom mask (#5510) · 19e1a5cf
      Hongxin Liu authored
      * [feature] refactor colo attention (#5462)
      
      * [extension] update api
      
      * [feature] add colo attention
      
      * [feature] update sdpa
      
      * [feature] update npu attention
      
      * [feature] update flash-attn
      
      * [test] add flash attn test
      
      * [test] update flash attn test
      
      * [shardformer] update modeling to fit colo attention (#5465)
      
      * [misc] refactor folder structure
      
      * [shardformer] update llama flash-attn
      
      * [shardformer] fix llama policy
      
      * [devops] update tensornvme install
      
      * [test] update llama test
      
      * [shardformer] update colo attn kernel dispatch
      
      * [shardformer] update blip2
      
      * [shardformer] update chatglm
      
      * [shardformer] update gpt2
      
      * [shardformer] update gptj
      
      * [shardformer] update opt
      
      * [shardformer] update vit
      
      * [shardformer] update colo attention mask prep
      
      * [shardformer] update whisper
      
      * [test] fix shardformer tests (#5514)
      
      * [test] fix shardformer tests
      
      * [test] fix shardformer tests
      19e1a5cf
  2. 26 Mar, 2024 1 commit
  3. 13 Mar, 2024 1 commit
    • Hongxin Liu's avatar
      [devops] fix compatibility (#5444) · f2e8b9ef
      Hongxin Liu authored
      * [devops] fix compatibility
      
      * [hotfix] update compatibility test on pr
      
      * [devops] fix compatibility
      
      * [devops] record duration during comp test
      
      * [test] decrease test duration
      
      * fix falcon
      f2e8b9ef
  4. 05 Mar, 2024 1 commit
  5. 04 Mar, 2024 1 commit
    • flybird11111's avatar
      [example]add gpt2 benchmark example script. (#5295) · 29695cf7
      flybird11111 authored
      
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [doc] fix typo in Colossal-LLaMA-2/README.md (#5247)
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed ddp test (#5254)
      
      * [ci] fixed ddp test
      
      * polish
      
      * fix typo in  applications/ColossalEval/README.md (#5250)
      
      * [ci] fix shardformer tests. (#5255)
      
      * fix ci
      
      fix
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [doc] fix doc typo (#5256)
      
      * [doc] fix annotation display
      
      * [doc] fix llama2 doc
      
      * [hotfix]: add pp sanity check and fix mbs arg (#5268)
      
      * fix: fix misleading mbs arg
      
      * feat: add pp sanity check
      
      * fix: fix 1f1b sanity check
      
      * [workflow] fixed incomplete bash command (#5272)
      
      * [workflow] fixed oom tests (#5275)
      
      * [workflow] fixed oom tests
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fix test_hybrid_parallel_plugin_checkpoint_io.py (#5276)
      
      * fix ci
      
      fix
      
      * fix test
      
      * revert: revert p2p
      
      * feat: add enable_metadata_cache option
      
      * revert: enable t5 tests
      
      * fix
      
      ---------
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      
      * [shardformer] hybridparallelplugin support gradients accumulation. (#5246)
      
      * support gradients acc
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * [hotfix] Fix ShardFormer test execution path when using sequence parallelism (#5230)
      
      * fix auto loading gpt2 tokenizer (#5279)
      
      * [doc] add llama2-13B disyplay (#5285)
      
      * Update README.md
      
      * fix 13b typo
      
      ---------
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      
      * fix llama pretrain (#5287)
      
      * fix
      
      * fix
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      fix
      
      * benchmark gpt2
      
      * fix
      
      fix
      
      fix
      
      fix
      
      * [workflow] fixed build CI (#5240)
      
      * [workflow] fixed build CI
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * polish
      
      * [ci] fixed booster test (#5251)
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * [ci] fixed booster test
      
      * fix
      
      fix
      
      * fix
      
      fix
      
      fix
      
      * fix
      
      * fix
      
      fix
      
      fix
      
      fix
      
      fix
      
      * fix
      
      * Update shardformer.py
      
      ---------
      Co-authored-by: default avatardigger yu <digger-yu@outlook.com>
      Co-authored-by: default avatarFrank Lee <somerlee.9@gmail.com>
      Co-authored-by: default avatarWenhao Chen <cwher@outlook.com>
      Co-authored-by: default avatarbinmakeswell <binmakeswell@gmail.com>
      Co-authored-by: default avatarZhongkai Zhao <kanezz620@gmail.com>
      Co-authored-by: default avatarMichelle <97082656+MichelleMa8@users.noreply.github.com>
      Co-authored-by: default avatarDesperado-Jia <502205863@qq.com>
      29695cf7
  6. 29 Feb, 2024 1 commit
  7. 27 Feb, 2024 1 commit
  8. 29 Jan, 2024 1 commit
  9. 25 Jan, 2024 1 commit
  10. 16 Jan, 2024 2 commits
  11. 11 Jan, 2024 1 commit
  12. 10 Jan, 2024 1 commit
  13. 09 Jan, 2024 1 commit
  14. 03 Jan, 2024 1 commit
  15. 28 Nov, 2023 1 commit
  16. 23 Nov, 2023 1 commit
    • YeAnbang's avatar
      [Feature] Add document retrieval QA (#5020) · e53e729d
      YeAnbang authored
      
      
      * add langchain
      
      * add langchain
      
      * Add files via upload
      
      * add langchain
      
      * fix style
      
      * fix style: remove extra space
      
      * add pytest; modified retriever
      
      * add pytest; modified retriever
      
      * add tests to build_on_pr.yml
      
      * fix build_on_pr.yml
      
      * fix build on pr; fix environ vars
      
      * seperate unit tests for colossalqa from build from pr
      
      * fix container setting; fix environ vars
      
      * commented dev code
      
      * add incremental update
      
      * remove stale code
      
      * fix style
      
      * change to sha3 224
      
      * fix retriever; fix style; add unit test for document loader
      
      * fix ci workflow config
      
      * fix ci workflow config
      
      * add set cuda visible device script in ci
      
      * fix doc string
      
      * fix style; update readme; refactored
      
      * add force log info
      
      * change build on pr, ignore colossalqa
      
      * fix docstring, captitalize all initial letters
      
      * fix indexing; fix text-splitter
      
      * remove debug code, update reference
      
      * reset previous commit
      
      * update LICENSE update README add key-value mode, fix bugs
      
      * add files back
      
      * revert force push
      
      * remove junk file
      
      * add test files
      
      * fix retriever bug, add intent classification
      
      * change conversation chain design
      
      * rewrite prompt and conversation chain
      
      * add ui v1
      
      * ui v1
      
      * fix atavar
      
      * add header
      
      * Refactor the RAG Code and support Pangu
      
      * Refactor the ColossalQA chain to Object-Oriented Programming and the UI demo.
      
      * resolved conversation. tested scripts under examples. web demo still buggy
      
      * fix ci tests
      
      * Some modifications to add ChatGPT api
      
      * modify llm.py and remove unnecessary files
      
      * Delete applications/ColossalQA/examples/ui/test_frontend_input.json
      
      * Remove OpenAI api key
      
      * add colossalqa
      
      * move files
      
      * move files
      
      * move files
      
      * move files
      
      * fix style
      
      * Add Readme and fix some bugs.
      
      * Add something to readme and modify some code
      
      * modify a directory name for clarity
      
      * remove redundant directory
      
      * Correct a type in  llm.py
      
      * fix AI prefix
      
      * fix test_memory.py
      
      * fix conversation
      
      * fix some erros and typos
      
      * Fix a missing import in RAG_ChatBot.py
      
      * add colossalcloud LLM wrapper, correct issues in code review
      
      ---------
      Co-authored-by: default avatarYeAnbang <anbangy2@outlook.com>
      Co-authored-by: default avatarOrion-Zheng <zheng_zian@u.nus.edu>
      Co-authored-by: default avatarZian(Andy) Zheng <62330719+Orion-Zheng@users.noreply.github.com>
      Co-authored-by: default avatarOrion-Zheng <zhengzian@u.nus.edu>
      e53e729d
  17. 08 Nov, 2023 1 commit
  18. 01 Nov, 2023 1 commit
  19. 27 Sep, 2023 1 commit
  20. 26 Sep, 2023 1 commit
  21. 20 Sep, 2023 1 commit
    • Wenhao Chen's avatar
      [chat]: update rm, add wandb and fix bugs (#4471) · 7b9b8644
      Wenhao Chen authored
      
      
      * feat: modify forward fn of critic and reward model
      
      * feat: modify calc_action_log_probs
      
      * to: add wandb in sft and rm trainer
      
      * feat: update train_sft
      
      * feat: update train_rm
      
      * style: modify type annotation and add warning
      
      * feat: pass tokenizer to ppo trainer
      
      * to: modify trainer base and maker base
      
      * feat: add wandb in ppo trainer
      
      * feat: pass tokenizer to generate
      
      * test: update generate fn tests
      
      * test: update train tests
      
      * fix: remove action_mask
      
      * feat: remove unused code
      
      * fix: fix wrong ignore_index
      
      * fix: fix mock tokenizer
      
      * chore: update requirements
      
      * revert: modify make_experience
      
      * fix: fix inference
      
      * fix: add padding side
      
      * style: modify _on_learn_batch_end
      
      * test: use mock tokenizer
      
      * fix: use bf16 to avoid overflow
      
      * fix: fix workflow
      
      * [chat] fix gemini strategy
      
      * [chat] fix
      
      * sync: update colossalai strategy
      
      * fix: fix args and model dtype
      
      * fix: fix checkpoint test
      
      * fix: fix requirements
      
      * fix: fix missing import and wrong arg
      
      * fix: temporarily skip gemini test in stage 3
      
      * style: apply pre-commit
      
      * fix: temporarily skip gemini test in stage 1&2
      
      ---------
      Co-authored-by: default avatarMingyan Jiang <1829166702@qq.com>
      7b9b8644
  22. 19 Sep, 2023 1 commit
  23. 18 Sep, 2023 1 commit
    • Hongxin Liu's avatar
      [legacy] clean up legacy code (#4743) · b5f9e37c
      Hongxin Liu authored
      * [legacy] remove outdated codes of pipeline (#4692)
      
      * [legacy] remove cli of benchmark and update optim (#4690)
      
      * [legacy] remove cli of benchmark and update optim
      
      * [doc] fix cli doc test
      
      * [legacy] fix engine clip grad norm
      
      * [legacy] remove outdated colo tensor (#4694)
      
      * [legacy] remove outdated colo tensor
      
      * [test] fix test import
      
      * [legacy] move outdated zero to legacy (#4696)
      
      * [legacy] clean up utils (#4700)
      
      * [legacy] clean up utils
      
      * [example] update examples
      
      * [legacy] clean up amp
      
      * [legacy] fix amp module
      
      * [legacy] clean up gpc (#4742)
      
      * [legacy] clean up context
      
      * [legacy] clean core, constants and global vars
      
      * [legacy] refactor initialize
      
      * [example] fix examples ci
      
      * [example] fix examples ci
      
      * [legacy] fix tests
      
      * [example] fix gpt example
      
      * [example] fix examples ci
      
      * [devops] fix ci installation
      
      * [example] fix examples ci
      b5f9e37c
  24. 11 Sep, 2023 1 commit
  25. 08 Sep, 2023 1 commit
  26. 01 Sep, 2023 1 commit
  27. 30 Aug, 2023 2 commits
  28. 16 Aug, 2023 1 commit
  29. 02 Aug, 2023 1 commit
    • Wenhao Chen's avatar
      [chat] fix bugs and add unit tests (#4213) · da4f7b85
      Wenhao Chen authored
      * style: rename replay buffer
      
      Experience replay is typically for off policy algorithms.
      Use this name in PPO maybe misleading.
      
      * fix: fix wrong zero2 default arg
      
      * test: update experience tests
      
      * style: rename zero_pad fn
      
      * fix: defer init in CycledDataLoader
      
      * test: add benchmark test
      
      * style: rename internal fn of generation
      
      * style: rename internal fn of lora
      
      * fix: remove unused loss fn
      
      * fix: remove unused utils fn
      
      * refactor: remove generate_with_actor fn
      
      * fix: fix type annotation
      
      * test: add models tests
      
      * fix: skip llama due to long execution time
      
      * style: modify dataset
      
      * style: apply formatter
      
      * perf: update reward dataset
      
      * fix: fix wrong IGNORE_INDEX in sft dataset
      
      * fix: remove DataCollatorForSupervisedDataset
      
      * test: add dataset tests
      
      * style: apply formatter
      
      * style: rename test_ci to test_train
      
      * feat: add llama in inference
      
      * test: add inference tests
      
      * test: change test scripts directory
      
      * fix: update ci
      
      * fix: fix typo
      
      * fix: skip llama due to oom
      
      * fix: fix file mod
      
      * style: apply formatter
      
      * refactor: remove duplicated llama_gptq
      
      * style: apply formatter
      
      * to: update rm test
      
      * feat: add tokenizer arg
      
      * feat: add download model script
      
      * test: update train tests
      
      * fix: modify gemini load and save pretrained
      
      * test: update checkpoint io test
      
      * to: modify nproc_per_node
      
      * fix: do not remove existing dir
      
      * fix: modify save path
      
      * test: add random choice
      
      * fix: fix sft path
      
      * fix: enlarge nproc_per_node to avoid oom
      
      * fix: add num_retry
      
      * fix: make lora config of rm and critic consistent
      
      * fix: add warning about lora weights
      
      * fix: skip some gpt2 tests
      
      * fix: remove grad ckpt in rm and critic due to errors
      
      * refactor: directly use Actor in train_sft
      
      * test: add more arguments
      
      * fix: disable grad ckpt when using lora
      
      * fix: fix save_pretrained and related tests
      
      * test: enable zero2 tests
      
      * revert: remove useless fn
      
      * style: polish code
      
      * test: modify test args
      da4f7b85
  30. 01 Aug, 2023 1 commit
  31. 21 Jul, 2023 1 commit
  32. 04 Jul, 2023 2 commits
    • Frank Lee's avatar
      [workflow] show test duration (#4159) · cc3cbe9f
      Frank Lee authored
      cc3cbe9f
    • Wenhao Chen's avatar
      [chat] use official transformers and fix some issues (#4117) · 3d8d5d0d
      Wenhao Chen authored
      * feat: remove on_learn_epoch fn as not used
      
      * revert: add _on_learn_epoch fn
      
      * feat: remove NaiveStrategy
      
      * test: update train_prompts tests
      
      * fix: remove prepare_llama_tokenizer_and_embedding
      
      * test: add lora arg
      
      * feat: remove roberta support in train_prompts due to runtime errs
      
      * feat: remove deberta & roberta in rm as not used
      
      * test: remove deberta and roberta tests
      
      * feat: remove deberta and roberta models as not used
      
      * fix: remove calls to roberta
      
      * fix: remove prepare_llama_tokenizer_and_embedding
      
      * chore: update transformers version
      
      * docs: update transformers version
      
      * fix: fix actor inference
      
      * fix: fix ci
      
      * feat: change llama pad token to unk
      
      * revert: revert ddp setup_distributed
      
      * fix: change llama pad token to unk
      
      * revert: undo unnecessary changes
      
      * fix: use pip to install transformers
      3d8d5d0d
  33. 28 Jun, 2023 1 commit
  34. 22 Jun, 2023 1 commit
  35. 19 Jun, 2023 1 commit
  36. 13 Jun, 2023 1 commit
  37. 12 Jun, 2023 1 commit