Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
TransformerEngine
Commits
eadb9886
Commit
eadb9886
authored
Aug 06, 2025
by
yuguo
Browse files
Merge branch 'develop_v2.5' of
http://10.16.6.30/dcutoolkit/deeplearing/TransformerEngine
parents
6aafc95e
eac0d49b
Changes
2
Show whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
7 additions
and
7 deletions
+7
-7
transformer_engine/common/gemm/cublaslt_gemm.cu
transformer_engine/common/gemm/cublaslt_gemm.cu
+6
-6
transformer_engine/pytorch/module/_common.py
transformer_engine/pytorch/module/_common.py
+1
-1
No files found.
transformer_engine/common/gemm/cublaslt_gemm.cu
View file @
eadb9886
...
...
@@ -1017,12 +1017,12 @@ void nvte_cublas_batchgemm(const NVTETensor A, const NVTETensor B, NVTETensor D,
int
math_sm_count
,
int
batch_count
,
cudaStream_t
stream
)
{
NVTE_API_CALL
(
nvte_cublas_batchgemm
);
using
namespace
transformer_engine
;
const
Tensor
*
inputA
=
reinterpret_cast
<
const
Tensor
*>
(
A
);
const
Tensor
*
inputB
=
reinterpret_cast
<
const
Tensor
*>
(
B
);
Tensor
*
outputD
=
reinterpret_cast
<
Tensor
*>
(
D
);
const
Tensor
*
biasTensor
=
reinterpret_cast
<
const
Tensor
*>
(
bias
);
Tensor
*
outputGelu
=
reinterpret_cast
<
Tensor
*>
(
pre_gelu_out
);
Tensor
*
wspace
=
reinterpret_cast
<
Tensor
*>
(
workspace
);
const
Tensor
*
inputA
=
convertNVTE
Tensor
Check
(
A
);
const
Tensor
*
inputB
=
convertNVTE
Tensor
Check
(
B
);
Tensor
*
outputD
=
convertNVTE
Tensor
(
D
);
const
Tensor
*
biasTensor
=
convertNVTE
Tensor
(
bias
);
Tensor
*
outputGelu
=
convertNVTE
Tensor
(
pre_gelu_out
);
Tensor
*
wspace
=
convertNVTE
Tensor
(
workspace
);
if
((
biasTensor
->
data
.
dptr
!=
nullptr
)
||
(
outputGelu
->
data
.
dptr
!=
nullptr
))
{
NVTE_ERROR
(
"MOE batchgemm not surpport bias or gelu."
);
}
...
...
transformer_engine/pytorch/module/_common.py
View file @
eadb9886
...
...
@@ -53,7 +53,7 @@ def apply_normalization(
normalization_func
=
_get_normalization_func
(
normalization
,
True
)
inputs
=
(
inputmat
,
ln_weight
)
if
ln_bias
is
None
else
(
inputmat
,
ln_weight
,
ln_bias
)
if
enable_lightop
and
(
ln_bias
is
None
)
and
normalization
==
"RMSNorm"
:
if
enable_lightop
and
(
ln_bias
is
None
)
and
normalization
==
"RMSNorm"
and
output_quantizer
is
None
and
(
output_dtype
is
torch
.
bfloat16
or
output_dtype
is
torch
.
float16
or
output_dtype
is
torch
.
float32
)
:
out
,
rsigma
=
rmsnorm_forward
(
inputmat
,
ln_weight
,
ln_out
,
eps
,
True
)
return
out
,
None
,
rsigma
else
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment