Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
TransformerEngine
Commits
9d26d942
Commit
9d26d942
authored
Aug 26, 2025
by
yuguo
Browse files
Merge branch 'develop_v2.5' of
http://10.16.6.30/dcutoolkit/deeplearing/TransformerEngine
parents
11bc1775
2e870ed9
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
2 additions
and
2 deletions
+2
-2
transformer_engine/pytorch/cpp_extensions/gemm.py
transformer_engine/pytorch/cpp_extensions/gemm.py
+1
-1
transformer_engine/pytorch/module/base.py
transformer_engine/pytorch/module/base.py
+1
-1
No files found.
transformer_engine/pytorch/cpp_extensions/gemm.py
View file @
9d26d942
...
...
@@ -570,7 +570,7 @@ def general_grouped_gemm(
dw
.
view
(
-
1
,
dw
.
size
(
-
1
)),
num_gemms
,
None
,
TE_DType
[
out_dtype
]
,
out_dtype
,
None
,
bias_dtype
,
gelu
,
...
...
transformer_engine/pytorch/module/base.py
View file @
9d26d942
...
...
@@ -1183,7 +1183,7 @@ class TransformerEngineBaseModule(torch.nn.Module, ABC):
):
grad_bias
=
grad_output
.
dequantize
().
view
(
-
1
,
grad_output
.
shape
[
-
1
]).
sum
(
dim
=
0
)
else
:
if
isinstance
(
quantizer
,
Float8BlockQuantizer
):
if
isinstance
(
quantizer
,
Float8BlockQuantizer
)
or
(
isinstance
(
quantizer
,
Float8CurrentScalingQuantizer
)
and
IS_HIP_EXTENSION
)
:
# unfuse bgrad for now until cast_transpose + dgrad calculation is ready for Float8BlockQuantizer.
grad_bias
=
grad_output
.
view
(
-
1
,
grad_output
.
shape
[
-
1
]).
sum
(
dim
=
0
)
else
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment