Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
TransformerEngine
Commits
aeceeac0
Commit
aeceeac0
authored
Jan 23, 2026
by
wenjh
Browse files
Merge branch 'develop_v2.10' into release_v2.10
parents
bd05b0dc
284d3f6f
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
2 additions
and
2 deletions
+2
-2
transformer_engine/common/__init__.py
transformer_engine/common/__init__.py
+1
-1
transformer_engine/pytorch/module/batched_linear.py
transformer_engine/pytorch/module/batched_linear.py
+1
-1
No files found.
transformer_engine/common/__init__.py
View file @
aeceeac0
...
@@ -392,10 +392,10 @@ if "NVTE_PROJECT_BUILDING" not in os.environ or bool(int(os.getenv("NVTE_RELEASE
...
@@ -392,10 +392,10 @@ if "NVTE_PROJECT_BUILDING" not in os.environ or bool(int(os.getenv("NVTE_RELEASE
_CURAND_LIB_CTYPES
=
_load_curand
()
_CURAND_LIB_CTYPES
=
_load_curand
()
_CUBLAS_LIB_CTYPES
=
_load_nvidia_cuda_library
(
"cublas"
)
_CUBLAS_LIB_CTYPES
=
_load_nvidia_cuda_library
(
"cublas"
)
_CUDART_LIB_CTYPES
=
_load_nvidia_cuda_library
(
"cuda_runtime"
)
_CUDART_LIB_CTYPES
=
_load_nvidia_cuda_library
(
"cuda_runtime"
)
_TE_LIB_CTYPES
=
_load_core_library
()
# Needed to find the correct headers for NVRTC kernels.
# Needed to find the correct headers for NVRTC kernels.
if
not
os
.
getenv
(
"NVTE_CUDA_INCLUDE_DIR"
)
and
_nvidia_cudart_include_dir
():
if
not
os
.
getenv
(
"NVTE_CUDA_INCLUDE_DIR"
)
and
_nvidia_cudart_include_dir
():
os
.
environ
[
"NVTE_CUDA_INCLUDE_DIR"
]
=
_nvidia_cudart_include_dir
()
os
.
environ
[
"NVTE_CUDA_INCLUDE_DIR"
]
=
_nvidia_cudart_include_dir
()
except
OSError
:
except
OSError
:
pass
pass
_TE_LIB_CTYPES
=
_load_core_library
()
transformer_engine/pytorch/module/batched_linear.py
View file @
aeceeac0
...
@@ -605,7 +605,7 @@ class BatchedLinear(TransformerEngineBaseModule):
...
@@ -605,7 +605,7 @@ class BatchedLinear(TransformerEngineBaseModule):
weight_tensors_fp8
=
[
None
]
*
int
(
self
.
num_gemms
)
weight_tensors_fp8
=
[
None
]
*
int
(
self
.
num_gemms
)
from
..cpu_offload
import
CPUOffloadEnabled
from
..cpu_offload
_v1
import
CPUOffloadEnabled
if
torch
.
is_grad_enabled
():
if
torch
.
is_grad_enabled
():
linear_fn
=
_BatchLinear
.
apply
linear_fn
=
_BatchLinear
.
apply
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment