Unverified Commit a7a69ca6 authored by Haowen Zheng's avatar Haowen Zheng Committed by GitHub
Browse files

Bump up FA to 2.8.3 (#2282)


Signed-off-by: default avatar将来 <jianglai.zhw@alibaba-inc.com>
Co-authored-by: default avatar将来 <jianglai.zhw@alibaba-inc.com>
Co-authored-by: default avatarKirthi Shankar Sivamani <ksivamani@nvidia.com>
parent bd380048
...@@ -18,10 +18,10 @@ sm_arch=`python3 -c "import torch; sm = torch.cuda.get_device_capability(0); pri ...@@ -18,10 +18,10 @@ sm_arch=`python3 -c "import torch; sm = torch.cuda.get_device_capability(0); pri
export FLASH_ATTN_CUDA_ARCHS=$sm_arch export FLASH_ATTN_CUDA_ARCHS=$sm_arch
if [ $sm_arch -gt 90 ] if [ $sm_arch -gt 90 ]
then then
FA_versions=(2.8.1) FA_versions=(2.8.3)
elif [ $sm_arch -eq 90 ] elif [ $sm_arch -eq 90 ]
then then
FA_versions=(2.7.3 2.8.1 3.0.0b1) FA_versions=(2.7.3 2.8.3 3.0.0b1)
fi fi
for fa_version in "${FA_versions[@]}" for fa_version in "${FA_versions[@]}"
......
...@@ -115,7 +115,7 @@ class FlashAttentionUtils: ...@@ -115,7 +115,7 @@ class FlashAttentionUtils:
version = PkgVersion("0") version = PkgVersion("0")
version_required = PkgVersion("2.1.1") version_required = PkgVersion("2.1.1")
version_required_blackwell = PkgVersion("2.7.3") version_required_blackwell = PkgVersion("2.7.3")
max_version = PkgVersion("2.8.1") max_version = PkgVersion("2.8.3")
v2_plus = False v2_plus = False
v2_1_plus = False v2_1_plus = False
v2_3_plus = False v2_3_plus = False
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment