@@ -5,6 +5,7 @@ This is a Triton implementation of the Flash Attention algorithm
(see: Dao et al., https://arxiv.org/pdf/2205.14135v2.pdf; Rabe and Staats https://arxiv.org/pdf/2112.05682v2.pdf; Triton https://github.com/openai/triton)
"""
importmath
importos
importsubprocess
...
...
@@ -36,17 +37,17 @@ except ImportError:
print('please install triton from https://github.com/openai/triton')