1. 16 Feb, 2022 1 commit
    • Sangkug Lym's avatar
      gradient accumulation fusion · 83b1e42f
      Sangkug Lym authored
      remove redundant linear layer class definition
      
      add fuse_gradient_accumulation attribute to weights for simple targetting
      
      reflect feedback and clean up the codes
      
      arg change
      83b1e42f
  2. 12 Jan, 2022 1 commit
  3. 11 Jan, 2022 1 commit
  4. 17 Dec, 2021 1 commit
  5. 19 Aug, 2021 1 commit
  6. 30 Jul, 2021 1 commit
    • Deepak Narayanan's avatar
      Support for pipeline parallelism in T5 model · 46c74b4c
      Deepak Narayanan authored
      - Accumulate encoder hidden state gradient to handle skip connection
      - Correctly compute the number of layers in encoder / decoder for T5 model
      - Ensure e weights are initialized the same way in embeddings
      - Synchronize embedding gradients across encoder and decoder for T5 model
      - Support for checkpoint loading and saving
      46c74b4c
  7. 09 Feb, 2021 2 commits
  8. 25 Jan, 2021 1 commit
  9. 04 Jan, 2021 1 commit
  10. 25 Dec, 2020 2 commits
  11. 22 Dec, 2020 1 commit
  12. 19 Dec, 2020 1 commit
  13. 12 Nov, 2020 2 commits
  14. 02 Sep, 2020 1 commit
  15. 06 Aug, 2020 1 commit
  16. 03 Aug, 2020 1 commit
  17. 26 Jul, 2020 2 commits
  18. 16 Apr, 2020 1 commit
  19. 29 Oct, 2019 1 commit
  20. 08 Oct, 2019 1 commit
  21. 30 Jul, 2019 1 commit