Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
b2435be6
Unverified
Commit
b2435be6
authored
Sep 15, 2025
by
b8zhong
Committed by
GitHub
Sep 15, 2025
Browse files
Cache the result of `is_blackwell` platform check (#10498)
parent
5fe39e85
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
3 additions
and
7 deletions
+3
-7
python/sglang/srt/layers/quantization/deep_gemm_wrapper/configurer.py
...g/srt/layers/quantization/deep_gemm_wrapper/configurer.py
+2
-7
python/sglang/srt/utils.py
python/sglang/srt/utils.py
+1
-0
No files found.
python/sglang/srt/layers/quantization/deep_gemm_wrapper/configurer.py
View file @
b2435be6
...
@@ -2,7 +2,7 @@ import logging
...
@@ -2,7 +2,7 @@ import logging
import
torch
import
torch
from
sglang.srt.utils
import
get_bool_env_var
,
get_device_sm
from
sglang.srt.utils
import
get_bool_env_var
,
get_device_sm
,
is_blackwell
logger
=
logging
.
getLogger
(
__name__
)
logger
=
logging
.
getLogger
(
__name__
)
...
@@ -21,12 +21,7 @@ def _compute_enable_deep_gemm():
...
@@ -21,12 +21,7 @@ def _compute_enable_deep_gemm():
return
get_bool_env_var
(
"SGL_ENABLE_JIT_DEEPGEMM"
,
default
=
"true"
)
return
get_bool_env_var
(
"SGL_ENABLE_JIT_DEEPGEMM"
,
default
=
"true"
)
def
_is_blackwell_arch
()
->
bool
:
major
,
minor
=
torch
.
cuda
.
get_device_capability
(
torch
.
cuda
.
current_device
())
return
major
==
10
ENABLE_JIT_DEEPGEMM
=
_compute_enable_deep_gemm
()
ENABLE_JIT_DEEPGEMM
=
_compute_enable_deep_gemm
()
DEEPGEMM_BLACKWELL
=
ENABLE_JIT_DEEPGEMM
and
_
is_blackwell
_arch
()
DEEPGEMM_BLACKWELL
=
ENABLE_JIT_DEEPGEMM
and
is_blackwell
()
DEEPGEMM_SCALE_UE8M0
=
DEEPGEMM_BLACKWELL
DEEPGEMM_SCALE_UE8M0
=
DEEPGEMM_BLACKWELL
python/sglang/srt/utils.py
View file @
b2435be6
...
@@ -167,6 +167,7 @@ is_ampere_with_cuda_12_3 = lambda: _check(8)
...
@@ -167,6 +167,7 @@ is_ampere_with_cuda_12_3 = lambda: _check(8)
is_hopper_with_cuda_12_3
=
lambda
:
_check
(
9
)
is_hopper_with_cuda_12_3
=
lambda
:
_check
(
9
)
@
lru_cache
(
maxsize
=
1
)
def
is_blackwell
():
def
is_blackwell
():
if
not
is_cuda
():
if
not
is_cuda
():
return
False
return
False
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment