feat: support flash attention 2 in qwen2 vl vision blocks (#2721)
* feat: support flash attention 2 in qwen2 vl vision blocks * fix: calc max_seqlen once and small refactors
Showing
Please register or sign in to comment
* feat: support flash attention 2 in qwen2 vl vision blocks * fix: calc max_seqlen once and small refactors