Add Megatron attention implementation for benchmarking
Showing
This diff is collapsed.
This diff is collapsed.
csrc/fused_softmax/setup.py
0 → 100644
flash_attn/fused_softmax.py
0 → 100644
Please register or sign in to comment