Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
TransformerEngine
Commits
2e870ed9
Commit
2e870ed9
authored
Aug 26, 2025
by
yuguo
Browse files
[DCU] fix
parent
059d92e2
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
2 additions
and
2 deletions
+2
-2
transformer_engine/pytorch/cpp_extensions/gemm.py
transformer_engine/pytorch/cpp_extensions/gemm.py
+1
-1
transformer_engine/pytorch/module/base.py
transformer_engine/pytorch/module/base.py
+1
-1
No files found.
transformer_engine/pytorch/cpp_extensions/gemm.py
View file @
2e870ed9
...
...
@@ -570,7 +570,7 @@ def general_grouped_gemm(
dw
.
view
(
-
1
,
dw
.
size
(
-
1
)),
num_gemms
,
None
,
TE_DType
[
out_dtype
]
,
out_dtype
,
None
,
bias_dtype
,
gelu
,
...
...
transformer_engine/pytorch/module/base.py
View file @
2e870ed9
...
...
@@ -1182,7 +1182,7 @@ class TransformerEngineBaseModule(torch.nn.Module, ABC):
):
grad_bias
=
grad_output
.
dequantize
().
view
(
-
1
,
grad_output
.
shape
[
-
1
]).
sum
(
dim
=
0
)
else
:
if
isinstance
(
quantizer
,
Float8BlockQuantizer
):
if
isinstance
(
quantizer
,
Float8BlockQuantizer
)
or
(
isinstance
(
quantizer
,
Float8CurrentScalingQuantizer
)
and
IS_HIP_EXTENSION
)
:
# unfuse bgrad for now until cast_transpose + dgrad calculation is ready for Float8BlockQuantizer.
grad_bias
=
grad_output
.
view
(
-
1
,
grad_output
.
shape
[
-
1
]).
sum
(
dim
=
0
)
else
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment