Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
c471d39e
Unverified
Commit
c471d39e
authored
May 20, 2025
by
fzyzcjy
Committed by
GitHub
May 19, 2025
Browse files
Support loading weights when physical experts are different from logical experts (#6386)
parent
d0443275
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
37 additions
and
1 deletion
+37
-1
python/sglang/srt/layers/moe/ep_moe/layer.py
python/sglang/srt/layers/moe/ep_moe/layer.py
+23
-0
python/sglang/srt/managers/expert_location.py
python/sglang/srt/managers/expert_location.py
+14
-1
No files found.
python/sglang/srt/layers/moe/ep_moe/layer.py
View file @
c471d39e
...
...
@@ -5,6 +5,7 @@ import torch
from
torch.nn
import
Module
from
sglang.srt.layers.quantization.deep_gemm
import
_ENABLE_JIT_DEEPGEMM
from
sglang.srt.managers.expert_location
import
get_global_expert_location_metadata
from
sglang.srt.managers.schedule_batch
import
global_server_args_dict
try
:
...
...
@@ -425,6 +426,28 @@ class EPMoE(torch.nn.Module):
weight_name
:
str
,
shard_id
:
str
,
expert_id
:
int
,
)
->
None
:
physical_expert_ids
=
(
get_global_expert_location_metadata
().
logical_to_all_physical
(
self
.
layer_id
,
expert_id
)
)
for
physical_expert_id
in
physical_expert_ids
:
self
.
_weight_loader_physical
(
param
=
param
,
loaded_weight
=
loaded_weight
,
weight_name
=
weight_name
,
shard_id
=
shard_id
,
expert_id
=
physical_expert_id
,
)
def
_weight_loader_physical
(
self
,
param
:
torch
.
nn
.
Parameter
,
loaded_weight
:
torch
.
Tensor
,
weight_name
:
str
,
shard_id
:
str
,
expert_id
:
int
,
)
->
None
:
if
expert_id
<
self
.
start_expert_id
or
expert_id
>
self
.
end_expert_id
:
return
...
...
python/sglang/srt/managers/expert_location.py
View file @
c471d39e
...
...
@@ -15,7 +15,7 @@ import json
import
logging
from
dataclasses
import
dataclass
from
pathlib
import
Path
from
typing
import
Optional
from
typing
import
List
,
Optional
import
torch
import
torch.distributed
...
...
@@ -163,6 +163,19 @@ class ExpertLocationMetadata:
logical_to_all_physical_map_num_valid
=
logical_to_all_physical_map_num_valid
,
)
# -------------------------------- usage ------------------------------------
def
logical_to_all_physical
(
self
,
layer_id
:
int
,
logical_expert_id
:
int
)
->
List
[
int
]:
return
[
physical_expert_id
for
physical_expert_id
in
self
.
logical_to_all_physical_map
[
layer_id
,
logical_expert_id
].
tolist
()
if
physical_expert_id
!=
-
1
]
_global_expert_location_metadata
:
Optional
[
ExpertLocationMetadata
]
=
None
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment