1. 06 Apr, 2023 4 commits
    • YY Lin's avatar
      [Chat]Add Peft support & fix the ptx bug (#3433) · 62f4e2eb
      YY Lin authored
      * Update ppo.py
      
      Fix the bug of fetching wrong batch data
      
      * Add peft model support in SFT and Prompts training
      
      In stage-1 and stage-3, the peft model supports are added. So the trained artifacts will be only a small lora additions instead of the whole bunch of files.
      
      * Delete test_prompts.txt
      
      * Delete test_pretrained.txt
      
      * Move the peft stuffs to a community folder.
      
      * Move the demo sft to community
      
      * delete dirty files
      
      * Add instructions to install peft using source
      
      * Remove Chinese comments
      
      * remove the Chinese comments
      62f4e2eb
    • Dr-Corgi's avatar
      [chat]fix save_model(#3377) · 73afb635
      Dr-Corgi authored
      The function save_model should be a part of PPOTrainer.
      73afb635
    • kingkingofall's avatar
      [chat]fix readme (#3429) · 57a3c4db
      kingkingofall authored
      * fix stage 2
      
      fix stage 2
      
      * add torch
      57a3c4db
    • Camille Zhong's avatar
      [Chat] fix the tokenizer "int too big to convert" error in SFT training (#3453) · 72cb4dd4
      Camille Zhong authored
      * Add RoBERTa for RLHF Stage 2 & 3 (test)
      
      RoBERTa for RLHF Stage 2 & 3 (still in testing)
      
      * Revert "Add RoBERTa for RLHF Stage 2 & 3 (test)"
      
      This reverts commit 06741d894dcbe958acd4e10d771f22275e20e368.
      
      * Add RoBERTa for RLHF stage 2 & 3
      
      1. add roberta folder under model folder
      2. add  roberta option in train_reward_model.py
      3. add some test in testci
      
      * Update test_ci.sh
      
      * Revert "Update test_ci.sh"
      
      This reverts commit 9c7352b81766f3177d31eeec0ec178a301df966a.
      
      * Add RoBERTa for RLHF Stage 2 & 3 (test)
      
      RoBERTa for RLHF Stage 2 & 3 (still in testing)
      
      * Revert "Add RoBERTa for RLHF Stage 2 & 3 (test)"
      
      This reverts commit 06741d894dcbe958acd4e10d771f22275e20e368.
      
      * Add RoBERTa for RLHF stage 2 & 3
      
      1. add roberta folder under model folder
      2. add  roberta option in train_reward_model.py
      3. add some test in testci
      
      * Update test_ci.sh
      
      * Revert "Update test_ci.sh"
      
      This reverts commit 9c7352b81766f3177d31eeec0ec178a301df966a.
      
      * update roberta with coati
      
      * chat ci update
      
      * Revert "chat ci update"
      
      This reverts commit 17ae7ae01fa752bd3289fc39069868fde99cf846.
      
      * [Chat] fix the tokenizer "int too big to convert" error in SFT training
      
      fix the tokenizer error during SFT training using Bloom and OPT
      72cb4dd4
  2. 05 Apr, 2023 1 commit
  3. 04 Apr, 2023 3 commits
  4. 03 Apr, 2023 1 commit
    • Camille Zhong's avatar
      [chatgpt] add pre-trained model RoBERTa for RLHF stage 2 & 3 (#3223) · 30412866
      Camille Zhong authored
      * Add RoBERTa for RLHF Stage 2 & 3 (test)
      
      RoBERTa for RLHF Stage 2 & 3 (still in testing)
      
      * Revert "Add RoBERTa for RLHF Stage 2 & 3 (test)"
      
      This reverts commit 06741d894dcbe958acd4e10d771f22275e20e368.
      
      * Add RoBERTa for RLHF stage 2 & 3
      
      1. add roberta folder under model folder
      2. add  roberta option in train_reward_model.py
      3. add some test in testci
      
      * add test for reward model training
      
      * Update test_ci.sh
      
      * Revert "Update test_ci.sh"
      
      This reverts commit 9c7352b81766f3177d31eeec0ec178a301df966a.
      
      * Add RoBERTa for RLHF Stage 2 & 3 (test)
      
      RoBERTa for RLHF Stage 2 & 3 (still in testing)
      
      * Revert "Add RoBERTa for RLHF Stage 2 & 3 (test)"
      
      This reverts commit 06741d894dcbe958acd4e10d771f22275e20e368.
      
      * Add RoBERTa for RLHF stage 2 & 3
      
      1. add roberta folder under model folder
      2. add  roberta option in train_reward_model.py
      3. add some test in testci
      
      * Update test_ci.sh
      
      * Revert "Update test_ci.sh"
      
      This reverts commit 9c7352b81766f3177d31eeec0ec178a301df966a.
      
      * update roberta with coati
      30412866
  5. 30 Mar, 2023 1 commit
  6. 29 Mar, 2023 6 commits
  7. 28 Mar, 2023 12 commits
  8. 24 Mar, 2023 4 commits
  9. 23 Mar, 2023 3 commits
  10. 22 Mar, 2023 3 commits
  11. 20 Mar, 2023 1 commit
    • BlueRum's avatar
      [chatgpt]Reward Model Training Process update (#3133) · 7548ca5a
      BlueRum authored
      * add normalize function to value_head in bloom rm
      
      * add normalization to value_function in gpt_rm
      
      * add normalization to value_head of opt_rm
      
      * add Anthropic/hh-rlhf dataset
      
      * Update __init__.py
      
      * Add LogExpLoss in RM training
      
      * Update __init__.py
      
      * update rm trainer to use acc as target
      
      * update example/train_rm
      
      * Update train_rm.sh
      
      * code style
      
      * Update README.md
      
      * Update README.md
      
      * add rm test to ci
      
      * fix tokenier
      
      * fix typo
      
      * change batchsize to avoid oom in ci
      
      * Update test_ci.sh
      7548ca5a
  12. 17 Mar, 2023 1 commit