1. 07 Feb, 2024 3 commits
    • Hongxin Liu's avatar
      [moe] support mixtral (#5309) · da39d21b
      Hongxin Liu authored
      * [moe] add mixtral block for single expert
      
      * [moe] mixtral block fwd support uneven ep
      
      * [moe] mixtral block bwd support uneven ep
      
      * [moe] add mixtral moe layer
      
      * [moe] simplify replace
      
      * [meo] support save sharded mixtral
      
      * [meo] support load sharded mixtral
      
      * [meo] support save sharded optim
      
      * [meo] integrate moe manager into plug
      
      * [meo] fix optimizer load
      
      * [meo] fix mixtral layer
      da39d21b
    • Hongxin Liu's avatar
      [moe] update capacity computing (#5253) · c904d2ae
      Hongxin Liu authored
      * [moe] top2 allow uneven input
      
      * [moe] update capacity computing
      
      * [moe] remove debug info
      
      * [moe] update capacity computing
      
      * [moe] update capacity computing
      c904d2ae
    • Xuanlei Zhao's avatar
      [moe] init mixtral impl · 7d8e0338
      Xuanlei Zhao authored
      7d8e0338
  2. 06 Feb, 2024 1 commit
  3. 04 Feb, 2024 1 commit
  4. 02 Feb, 2024 1 commit
  5. 01 Feb, 2024 2 commits
  6. 31 Jan, 2024 1 commit
  7. 30 Jan, 2024 2 commits
  8. 29 Jan, 2024 6 commits
  9. 25 Jan, 2024 3 commits
  10. 22 Jan, 2024 2 commits
  11. 19 Jan, 2024 2 commits
  12. 18 Jan, 2024 3 commits
  13. 17 Jan, 2024 3 commits
  14. 16 Jan, 2024 2 commits
  15. 15 Jan, 2024 1 commit
  16. 11 Jan, 2024 5 commits
  17. 10 Jan, 2024 2 commits