Unverified Commit 6c526794 authored by vcherepanov-nv's avatar vcherepanov-nv Committed by GitHub
Browse files

Bump up FA to 2.8.1 (#1949)


Signed-off-by: default avatarVladimir Cherepanov <vcherepanov@nvidia.com>
parent e7251f93
......@@ -18,10 +18,10 @@ sm_arch=`python3 -c "import torch; sm = torch.cuda.get_device_capability(0); pri
export FLASH_ATTN_CUDA_ARCHS=$sm_arch
if [ $sm_arch -gt 90 ]
then
FA_versions=(2.7.3)
FA_versions=(2.8.1)
elif [ $sm_arch -eq 90 ]
then
FA_versions=(2.5.7 2.7.3 3.0.0b1)
FA_versions=(2.7.3 2.8.1 3.0.0b1)
fi
for fa_version in "${FA_versions[@]}"
......
......@@ -105,7 +105,7 @@ class FlashAttentionUtils:
version = PkgVersion("0")
version_required = PkgVersion("2.1.1")
version_required_blackwell = PkgVersion("2.7.3")
max_version = PkgVersion("2.7.4.post1")
max_version = PkgVersion("2.8.1")
v2_plus = False
v2_1_plus = False
v2_3_plus = False
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment