Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
01d47a27
"vscode:/vscode.git/clone" did not exist on "fc5616c8d0d722263d55a760dbe8b7641d88bca1"
Unverified
Commit
01d47a27
authored
Aug 20, 2025
by
chenxu140
Committed by
GitHub
Aug 19, 2025
Browse files
[Bugfix] fix kv buffer register & dp attention & deepepmoe (#9327)
parent
ecc9f3e4
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
3 additions
and
5 deletions
+3
-5
python/sglang/srt/disaggregation/ascend/conn.py
python/sglang/srt/disaggregation/ascend/conn.py
+1
-3
python/sglang/srt/layers/dp_attention.py
python/sglang/srt/layers/dp_attention.py
+1
-1
python/sglang/srt/layers/moe/ep_moe/layer.py
python/sglang/srt/layers/moe/ep_moe/layer.py
+1
-1
No files found.
python/sglang/srt/disaggregation/ascend/conn.py
View file @
01d47a27
...
...
@@ -23,9 +23,7 @@ class AscendKVManager(MooncakeKVManager):
)
def
register_buffer_to_engine
(
self
):
self
.
engine
.
register
(
self
.
kv_args
.
kv_data_ptrs
[
0
],
sum
(
self
.
kv_args
.
kv_data_lens
)
)
self
.
engine
.
batch_register
(
self
.
kv_args
.
kv_data_ptrs
,
self
.
kv_args
.
kv_data_lens
)
# The Ascend backend optimize batch registration for small memory blocks.
self
.
engine
.
batch_register
(
self
.
kv_args
.
aux_data_ptrs
,
self
.
kv_args
.
aux_data_lens
...
...
python/sglang/srt/layers/dp_attention.py
View file @
01d47a27
...
...
@@ -234,7 +234,7 @@ def initialize_dp_attention(
_DpGatheredBufferWrapper
.
set_metadata
(
hidden_size
=
model_config
.
hidden_size
,
dtype
=
model_config
.
dtype
,
device
=
torch
.
device
(
"cuda"
),
device
=
torch
.
device
(
server_args
.
device
),
)
...
...
python/sglang/srt/layers/moe/ep_moe/layer.py
View file @
01d47a27
...
...
@@ -736,7 +736,7 @@ class DeepEPMoE(EPMoE):
assert
isinstance
(
dispatch_output
,
AscendDeepEPLLOutput
)
hidden_states
,
topk_idx
,
topk_weights
,
_
,
seg_indptr
,
_
=
dispatch_output
assert
self
.
quant_method
is
not
None
assert
self
.
activation
==
"silu"
assert
self
.
moe_runner_config
.
activation
==
"silu"
# NOTE: Ascend's Dispatch & Combine does not support FP16
output_dtype
=
torch
.
bfloat16
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment