1. 17 Apr, 2023 8 commits
    • digger-yu's avatar
      [doc] Update .github/workflows/README.md (#3577) · 6e7e43c6
      digger-yu authored
      Optimization Code
      I think there were two extra $ entered here, which have been deleted
      6e7e43c6
    • Fazzie-Maqianli's avatar
      6b1a39b1
    • binmakeswell's avatar
      [chat] update reward model sh (#3578) · cc1eec2f
      binmakeswell authored
      cc1eec2f
    • csric's avatar
      [chatgpt] Detached PPO Training (#3195) · e3551443
      csric authored
      
      
      * run the base
      
      * working on dist ppo
      
      * sync
      
      * detached trainer
      
      * update detached trainer. no maker update function
      
      * facing init problem
      
      * 1 maker 1 trainer detached run. but no model update
      
      * facing cuda problem
      
      * fix save functions
      
      * verified maker update
      
      * nothing
      
      * add ignore
      
      * analyize loss issue
      
      * remove some debug codes
      
      * facing 2m1t stuck issue
      
      * 2m1t verified
      
      * do not use torchrun
      
      * working on 2m2t
      
      * working on 2m2t
      
      * initialize strategy in ray actor env
      
      * facing actor's init order issue
      
      * facing ddp model update issue (need unwarp ddp)
      
      * unwrap ddp actor
      
      * checking 1m2t stuck problem
      
      * nothing
      
      * set timeout for trainer choosing. It solves the stuck problem!
      
      * delete some debug output
      
      * rename to sync with upstream
      
      * rename to sync with upstream
      
      * coati rename
      
      * nothing
      
      * I am going to detach the replaybuffer from trainer and make it a Ray Actor. Two benefits: 1. support TP trainer. 2. asynchronized buffer operations
      
      * experience_maker_holder performs target-revolving _send_experience() instead of length comparison.
      
      * move code to ray subfolder
      
      * working on pipeline inference
      
      * apply comments
      
      ---------
      Co-authored-by: default avatarcsric <richcsr256@gmail.com>
      e3551443
    • YH's avatar
      Add docstr for zero3 chunk search utils (#3572) · d329c294
      YH authored
      d329c294
    • digger-yu's avatar
      [doc] Update 1D_tensor_parallel.md (#3573) · 9edeadfb
      digger-yu authored
      Display format optimization , same as fix#3562
      Simultaneous modification of en version
      9edeadfb
    • Hongxin Liu's avatar
      [misc] add verbose arg for zero and op builder (#3552) · 173dad05
      Hongxin Liu authored
      * [misc] add print verbose
      
      * [gemini] add print verbose
      
      * [zero] add print verbose for low level
      
      * [misc] add print verbose for op builder
      173dad05
    • Hongxin Liu's avatar
      [lazyinit] fix clone and deepcopy (#3553) · 4341f5e8
      Hongxin Liu authored
      4341f5e8
  2. 14 Apr, 2023 2 commits
    • digger-yu's avatar
      [doc] Update 1D_tensor_parallel.md (#3563) · 1c7734bc
      digger-yu authored
      Display format optimization, fix bug#3562
      Specific changes
      1. "This is called a column-parallel fashion" Translate to Chinese
      2. use the ```math code block syntax to display a math expression as a block, No modification of formula content
      
      Please check that the math formula is displayed correctly
      If OK, I will change the format of the English version of the formula in parallel
      1c7734bc
    • binmakeswell's avatar
      f1b3d60c
  3. 13 Apr, 2023 4 commits
  4. 12 Apr, 2023 5 commits
  5. 11 Apr, 2023 1 commit
  6. 10 Apr, 2023 5 commits
  7. 07 Apr, 2023 3 commits
  8. 06 Apr, 2023 12 commits