Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
cd641a99
Unverified
Commit
cd641a99
authored
Sep 24, 2025
by
GuoweiWangU
Committed by
GitHub
Sep 24, 2025
Browse files
fix bailing_moe with enable_dp_attention (#10860)
parent
71f24ef8
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
2 additions
and
2 deletions
+2
-2
python/sglang/srt/models/bailing_moe.py
python/sglang/srt/models/bailing_moe.py
+2
-2
No files found.
python/sglang/srt/models/bailing_moe.py
View file @
cd641a99
...
...
@@ -45,12 +45,12 @@ from sglang.srt.layers.dp_attention import (
get_attention_dp_size
,
get_attention_tp_rank
,
get_attention_tp_size
,
is_dp_attention_enabled
,
)
from
sglang.srt.layers.layernorm
import
RMSNorm
from
sglang.srt.layers.linear
import
(
MergedColumnParallelLinear
,
QKVParallelLinear
,
ReplicatedLinear
,
RowParallelLinear
,
)
from
sglang.srt.layers.logits_processor
import
LogitsProcessor
...
...
@@ -702,7 +702,7 @@ class BailingMoEModel(nn.Module):
self
.
embed_dim
,
quant_config
=
quant_config
,
prefix
=
add_prefix
(
"word_embeddings"
,
prefix
),
use_attn_tp_group
=
global_server_args_dict
[
"enable_dp_lm_head"
]
,
enable_tp
=
not
is_dp_attention_enabled
()
,
)
else
:
self
.
word_embeddings
=
PPMissingLayer
()
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment