Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
evt_fugx1
dcu_megatron
Commits
7a088067
"tests/distributed/synced_batchnorm/unit_test.sh" did not exist on "e12c1ec300c6b7369e17ba733996eacefae462a9"
Commit
7a088067
authored
May 26, 2025
by
dongcl
Browse files
1f1b overlap only supports MoEAlltoAllTokenDispatcher
parent
6dcd0fb8
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
4 deletions
+6
-4
dcu_megatron/core/transformer/transformer_layer.py
dcu_megatron/core/transformer/transformer_layer.py
+6
-4
No files found.
dcu_megatron/core/transformer/transformer_layer.py
View file @
7a088067
...
...
@@ -12,8 +12,7 @@ from megatron.core.utils import (
)
from
megatron.core.transformer.moe.moe_layer
import
MoELayer
from
megatron.core.transformer.transformer_layer
import
TransformerLayer
as
MegatronCoreTransformerLayer
from
dcu_megatron.core.transformer.utils
import
SubmoduleCallables
,
TransformerLayerSubmoduleCallables
from
megatron.core.transformer.moe.token_dispatcher
import
MoEAlltoAllTokenDispatcher
class
TransformerLayer
(
MegatronCoreTransformerLayer
):
...
...
@@ -34,7 +33,10 @@ class TransformerLayer(MegatronCoreTransformerLayer):
inference_params
:
Optional
[
Any
]
=
None
,
):
if
not
isinstance
(
self
.
mlp
,
MoELayer
):
if
(
not
isinstance
(
self
.
mlp
,
MoELayer
)
or
not
isinstance
(
self
.
mlp
.
token_dispatcher
,
MoEAlltoAllTokenDispatcher
)
):
return
super
().
forward
(
hidden_states
=
hidden_states
,
context
=
context
,
...
...
@@ -55,7 +57,7 @@ class TransformerLayer(MegatronCoreTransformerLayer):
pre_mlp_layernorm_output
,
tokens_per_expert
,
permutated_local_input_tokens
,
probs
,
_
,
)
=
self
.
_submodule_attention_router_compound_forward
(
hidden_states
,
attention_mask
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment