Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
d61615fe
"...graphsage/git@developer.sourcefind.cn:OpenDAS/dgl.git" did not exist on "5152a87982c3cddd8429ea0b2e777830fa48ae5c"
Unverified
Commit
d61615fe
authored
Oct 02, 2025
by
fzyzcjy
Committed by
GitHub
Oct 02, 2025
Browse files
Tiny fix missing alt stream in nextn layer (#10768)
parent
b1ccaf01
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
1 deletion
+6
-1
python/sglang/srt/models/deepseek_nextn.py
python/sglang/srt/models/deepseek_nextn.py
+6
-1
No files found.
python/sglang/srt/models/deepseek_nextn.py
View file @
d61615fe
...
@@ -33,11 +33,14 @@ from sglang.srt.layers.vocab_parallel_embedding import (
...
@@ -33,11 +33,14 @@ from sglang.srt.layers.vocab_parallel_embedding import (
from
sglang.srt.managers.schedule_batch
import
global_server_args_dict
from
sglang.srt.managers.schedule_batch
import
global_server_args_dict
from
sglang.srt.model_executor.forward_batch_info
import
ForwardBatch
from
sglang.srt.model_executor.forward_batch_info
import
ForwardBatch
from
sglang.srt.models.deepseek_v2
import
DeepseekV2DecoderLayer
,
DeepseekV3ForCausalLM
from
sglang.srt.models.deepseek_v2
import
DeepseekV2DecoderLayer
,
DeepseekV3ForCausalLM
from
sglang.srt.utils
import
BumpAllocator
,
add_prefix
from
sglang.srt.utils
import
BumpAllocator
,
add_prefix
,
is_cuda
logger
=
logging
.
getLogger
(
__name__
)
logger
=
logging
.
getLogger
(
__name__
)
_is_cuda
=
is_cuda
()
class
DeepseekModelNextN
(
nn
.
Module
):
class
DeepseekModelNextN
(
nn
.
Module
):
def
__init__
(
def
__init__
(
self
,
self
,
...
@@ -66,12 +69,14 @@ class DeepseekModelNextN(nn.Module):
...
@@ -66,12 +69,14 @@ class DeepseekModelNextN(nn.Module):
self
.
eh_proj
=
nn
.
Linear
(
2
*
config
.
hidden_size
,
config
.
hidden_size
,
bias
=
False
)
self
.
eh_proj
=
nn
.
Linear
(
2
*
config
.
hidden_size
,
config
.
hidden_size
,
bias
=
False
)
self
.
alt_stream
=
torch
.
cuda
.
Stream
()
if
_is_cuda
else
None
self
.
decoder
=
DeepseekV2DecoderLayer
(
self
.
decoder
=
DeepseekV2DecoderLayer
(
config
,
config
,
0
,
0
,
quant_config
=
quant_config
,
quant_config
=
quant_config
,
is_nextn
=
True
,
is_nextn
=
True
,
prefix
=
add_prefix
(
"decoder"
,
prefix
),
prefix
=
add_prefix
(
"decoder"
,
prefix
),
alt_stream
=
self
.
alt_stream
,
)
)
self
.
shared_head
=
nn
.
Module
()
self
.
shared_head
=
nn
.
Module
()
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment