[npu] support triangle attention for llama (#5130)
* update fused attn * update spda * tri attn * update triangle * import * fix * fix
Showing
Please register or sign in to comment
* update fused attn * update spda * tri attn * update triangle * import * fix * fix