Commit 9200e43a authored by Tri Dao's avatar Tri Dao
Browse files

Remove FA's check for headdim <= 128

parent d693034e
......@@ -386,8 +386,6 @@ class ParallelAttention(MegatronModule):
'self-attention for now')
assert self.attn_mask_type == AttnMaskType.causal, ('FlashAttention code path only '
'supports causal mask for now')
headdim = args.hidden_size / args.num_attention_heads
assert headdim <= 128, 'FlashAttention only supports head dimension at most 128'
if rearrange is None:
raise ImportError('einops is not installed, please install with pip install einops')
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment