Unverified Commit a7a69ca6 authored by Haowen Zheng's avatar Haowen Zheng Committed by GitHub
Browse files

Bump up FA to 2.8.3 (#2282)


Signed-off-by: default avatar将来 <jianglai.zhw@alibaba-inc.com>
Co-authored-by: default avatar将来 <jianglai.zhw@alibaba-inc.com>
Co-authored-by: default avatarKirthi Shankar Sivamani <ksivamani@nvidia.com>
parent bd380048
......@@ -18,10 +18,10 @@ sm_arch=`python3 -c "import torch; sm = torch.cuda.get_device_capability(0); pri
export FLASH_ATTN_CUDA_ARCHS=$sm_arch
if [ $sm_arch -gt 90 ]
then
FA_versions=(2.8.1)
FA_versions=(2.8.3)
elif [ $sm_arch -eq 90 ]
then
FA_versions=(2.7.3 2.8.1 3.0.0b1)
FA_versions=(2.7.3 2.8.3 3.0.0b1)
fi
for fa_version in "${FA_versions[@]}"
......
......@@ -115,7 +115,7 @@ class FlashAttentionUtils:
version = PkgVersion("0")
version_required = PkgVersion("2.1.1")
version_required_blackwell = PkgVersion("2.7.3")
max_version = PkgVersion("2.8.1")
max_version = PkgVersion("2.8.3")
v2_plus = False
v2_1_plus = False
v2_3_plus = False
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment