Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
TransformerEngine
Commits
99e60246
Commit
99e60246
authored
Dec 03, 2025
by
wenjh
Browse files
Make release_v2.9 compile pass
parent
cbb14a5f
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
3 additions
and
3 deletions
+3
-3
transformer_engine/common/include/transformer_engine/gemm.h
transformer_engine/common/include/transformer_engine/gemm.h
+1
-1
transformer_engine/common/swizzle/swizzle_block_scaling.cu
transformer_engine/common/swizzle/swizzle_block_scaling.cu
+1
-1
transformer_engine/pytorch/csrc/util.cpp
transformer_engine/pytorch/csrc/util.cpp
+1
-1
No files found.
transformer_engine/common/include/transformer_engine/gemm.h
View file @
99e60246
...
@@ -130,7 +130,7 @@ void nvte_cublas_gemm(const NVTETensor A, const NVTETensor B, NVTETensor D, cons
...
@@ -130,7 +130,7 @@ void nvte_cublas_gemm(const NVTETensor A, const NVTETensor B, NVTETensor D, cons
*/
*/
void
nvte_cublas_gemm_v2
(
int
transa
,
int
transb
,
const
float
*
alpha
,
const
NVTETensor
A
,
void
nvte_cublas_gemm_v2
(
int
transa
,
int
transb
,
const
float
*
alpha
,
const
NVTETensor
A
,
const
NVTETensor
B
,
const
float
*
beta
,
const
NVTETensor
C
,
NVTETensor
D
,
const
NVTETensor
B
,
const
float
*
beta
,
const
NVTETensor
C
,
NVTETensor
D
,
NVTETensor
workspace
,
NVTEMatmulConfig
config
,
cudaStream_t
stream
,
bool
nvte_use_hipblaslt
,
bool
nvte_use_rocblas
,
int
compute_stream_offset
);
NVTETensor
workspace
,
NVTEMatmulConfig
config
,
cudaStream_t
stream
,
bool
nvte_use_hipblaslt
=
0
,
bool
nvte_use_rocblas
=
0
,
int
compute_stream_offset
=
0
);
/*! \brief Compute matrix multiplication of 2 matrices, potentially fused with other operations,
/*! \brief Compute matrix multiplication of 2 matrices, potentially fused with other operations,
* allowing for using a scaling factor for the GEMM result and the accumulation input (deprecated)
* allowing for using a scaling factor for the GEMM result and the accumulation input (deprecated)
...
...
transformer_engine/common/swizzle/swizzle_block_scaling.cu
View file @
99e60246
...
@@ -14,7 +14,7 @@
...
@@ -14,7 +14,7 @@
#include "../util/logging.h"
#include "../util/logging.h"
#include "transformer_engine/transformer_engine.h"
#include "transformer_engine/transformer_engine.h"
#ifdef __HIP_PLATFORM_AMD__
#if
n
def __HIP_PLATFORM_AMD__
namespace
transformer_engine
{
namespace
transformer_engine
{
namespace
{
namespace
{
constexpr
uint32_t
WARP_SIZE
=
32
;
constexpr
uint32_t
WARP_SIZE
=
32
;
...
...
transformer_engine/pytorch/csrc/util.cpp
View file @
99e60246
...
@@ -228,7 +228,7 @@ at::Tensor convert_block_scaling_to_mxfp8_tensor(transformer_engine::TensorWrapp
...
@@ -228,7 +228,7 @@ at::Tensor convert_block_scaling_to_mxfp8_tensor(transformer_engine::TensorWrapp
// Allocate memory for swizzled mxfp8 scaling factors
// Allocate memory for swizzled mxfp8 scaling factors
const
auto
options
=
at
::
TensorOptions
().
dtype
(
torch
::
kByte
).
device
(
torch
::
kCUDA
);
const
auto
options
=
at
::
TensorOptions
().
dtype
(
torch
::
kByte
).
device
(
torch
::
kCUDA
);
at
::
Tensor
swizzled_scale_inv
=
at
::
empty
(
at
::
Tensor
swizzled_scale_inv
=
at
::
empty
(
std
::
vector
<
int64_t
>
{
swizzled_scale_inv_first_dim
,
swizzled_scale_inv_last_dim
},
options
);
std
::
vector
<
int64_t
>
{
static_cast
<
int64_t
>
(
swizzled_scale_inv_first_dim
)
,
static_cast
<
int64_t
>
(
swizzled_scale_inv_last_dim
)
},
options
);
// Set rowwise scaling factors on output
// Set rowwise scaling factors on output
void
*
const
swizzled_scale_inv_dptr
=
getDataPtr
(
swizzled_scale_inv
,
0
);
void
*
const
swizzled_scale_inv_dptr
=
getDataPtr
(
swizzled_scale_inv
,
0
);
NVTEShape
swizzled_scale_inv_shape
{};
NVTEShape
swizzled_scale_inv_shape
{};
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment