- 12 Dec, 2025 1 commit
-
-
Lei Wang authored
-
- 17 Nov, 2025 1 commit
-
-
Tong WU authored
[Enhancement] Keep max score attention across blocks in FlashAttention for better numerical stablity (#1269) * Implement max score retention across blocks in FlashAttention for improved stability * fix manual pipeline parameters * Update examples/flash_attention/example_gqa_fwd_varlen.py Co-authored-by:
coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * fix typo * more * fix a previous typo --------- Co-authored-by:
coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com>
-
- 21 Oct, 2025 1 commit
-
-
Tong WU authored
* [Cleanup] Remove `tilelang.disable_cache()` calls from example scripts * lint * lint
-
- 05 Sep, 2025 1 commit
-
-
Wenhao Xie authored
* fix * lint
-
- 28 Aug, 2025 1 commit
-
-
Wenhao Xie authored
* upd sparse attn * lint * rename * update test file * update benchmark * lint * update benchmark
-