- 25 Jan, 2024 1 commit
-
-
Frank Lee authored
* [feat] refactored extension module * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish
-
- 08 Jan, 2024 1 commit
-
-
Xuanlei Zhao authored
* update extension * update cpu adam * update is * add doc for cpu adam * update kernel * update commit * update flash * update memory efficient * update flash attn * update flash attention loader * update api * fix * update doc * update example time limit * reverse change * fix doc * remove useless kernel * fix * not use warning * update * update
-
- 19 Sep, 2023 1 commit
-
-
Hongxin Liu authored
* [misc] update pre-commit * [misc] run pre-commit * [misc] remove useless configuration files * [misc] ignore cuda for clang-format
-
- 15 Aug, 2023 1 commit
-
-
flybird1111 authored
* cherry-pick flash attention 2 cherry-pick flash attention 2 * [shardformer] update shardformer to use flash attention 2 [shardformer] update shardformer to use flash attention 2, fix [shardformer] update shardformer to use flash attention 2, fix [shardformer] update shardformer to use flash attention 2, fix
-
- 04 Aug, 2023 1 commit
-
-
flybird1111 authored
Improved ColoAttention interface to support flash attention 2. Solved #4322
-
- 03 Feb, 2023 1 commit
-
-
Frank Lee authored
* [kernel] fixed repeated loading of kernels * polish code * polish code
-
- 15 Nov, 2022 1 commit
-
-
zbian authored
-
- 13 Jan, 2022 1 commit
-
-
ver217 authored
-
- 21 Dec, 2021 1 commit
-
-
shenggan authored
-