Unverified Commit d90749d3 authored by Atream's avatar Atream Committed by GitHub
Browse files

Update triton_attention.py

parent 1548c992
# Adapted from
# https://github.com/sgl-project/sglang/blob/9f635ea50de920aa507f486daafba26a5b837574/python/sglang/srt/layers/attention/triton_ops/decode_attention.py
# which was originally adapted from
# https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage1.py
# https://github.com/ModelTC/lightllm/blob/96353e868a840db4d103138caf15ed9dbea8c186/lightllm/models/deepseek2/triton_kernel/gqa_flash_decoding_stage2.py
import triton import triton
import triton.language as tl import triton.language as tl
...@@ -376,4 +382,4 @@ def decode_attention_fwd_grouped( ...@@ -376,4 +382,4 @@ def decode_attention_fwd_grouped(
) )
_decode_softmax_reducev_fwd(attn_logits, q, o, v_buffer, b_seq_len, _decode_softmax_reducev_fwd(attn_logits, q, o, v_buffer, b_seq_len,
num_kv_splits) num_kv_splits)
\ No newline at end of file
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment