Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
44cb0607
Unverified
Commit
44cb0607
authored
Oct 09, 2025
by
Yineng Zhang
Committed by
GitHub
Oct 09, 2025
Browse files
chore: upgrade flashinfer 0.4.0 (#11364)
parent
88bb627d
Changes
5
Hide whitespace changes
Inline
Side-by-side
Showing
5 changed files
with
6 additions
and
4 deletions
+6
-4
python/pyproject.toml
python/pyproject.toml
+1
-1
python/pyproject_other.toml
python/pyproject_other.toml
+1
-1
python/sglang/srt/entrypoints/engine.py
python/sglang/srt/entrypoints/engine.py
+1
-1
python/sglang/srt/layers/attention/flashinfer_mla_backend.py
python/sglang/srt/layers/attention/flashinfer_mla_backend.py
+1
-1
scripts/ci/ci_install_dependency.sh
scripts/ci/ci_install_dependency.sh
+2
-0
No files found.
python/pyproject.toml
View file @
44cb0607
...
@@ -24,7 +24,7 @@ dependencies = [
...
@@ -24,7 +24,7 @@ dependencies = [
"datasets"
,
"datasets"
,
"einops"
,
"einops"
,
"fastapi"
,
"fastapi"
,
"flashinfer_python==0.4.0
rc3
"
,
"flashinfer_python==0.4.0"
,
"hf_transfer"
,
"hf_transfer"
,
"huggingface_hub"
,
"huggingface_hub"
,
"interegular"
,
"interegular"
,
...
...
python/pyproject_other.toml
View file @
44cb0607
...
@@ -70,7 +70,7 @@ srt = [
...
@@ -70,7 +70,7 @@ srt = [
"torchaudio==2.8.0"
,
"torchaudio==2.8.0"
,
"torchvision"
,
"torchvision"
,
"cuda-python"
,
"cuda-python"
,
"flashinfer_python==0.4.0
rc3
"
,
"flashinfer_python==0.4.0"
,
]
]
# HIP (Heterogeneous-computing Interface for Portability) for AMD
# HIP (Heterogeneous-computing Interface for Portability) for AMD
...
...
python/sglang/srt/entrypoints/engine.py
View file @
44cb0607
...
@@ -703,7 +703,7 @@ def _set_envs_and_config(server_args: ServerArgs):
...
@@ -703,7 +703,7 @@ def _set_envs_and_config(server_args: ServerArgs):
if
server_args
.
attention_backend
==
"flashinfer"
:
if
server_args
.
attention_backend
==
"flashinfer"
:
assert_pkg_version
(
assert_pkg_version
(
"flashinfer_python"
,
"flashinfer_python"
,
"0.4.0
rc3
"
,
"0.4.0"
,
"Please uninstall the old version and "
"Please uninstall the old version and "
"reinstall the latest version by following the instructions "
"reinstall the latest version by following the instructions "
"at https://docs.flashinfer.ai/installation.html."
,
"at https://docs.flashinfer.ai/installation.html."
,
...
...
python/sglang/srt/layers/attention/flashinfer_mla_backend.py
View file @
44cb0607
...
@@ -1060,7 +1060,7 @@ def fast_mla_decode_plan(
...
@@ -1060,7 +1060,7 @@ def fast_mla_decode_plan(
try
:
try
:
# Standard version with just the required arguments (no use_profiler)
# Standard version with just the required arguments (no use_profiler)
self
.
_cached_module
.
plan
.
default
(
self
.
_cached_module
.
plan
(
self
.
_float_workspace_buffer
,
self
.
_float_workspace_buffer
,
self
.
_int_workspace_buffer
,
self
.
_int_workspace_buffer
,
self
.
_pin_memory_int_workspace_buffer
,
self
.
_pin_memory_int_workspace_buffer
,
...
...
scripts/ci/ci_install_dependency.sh
View file @
44cb0607
...
@@ -74,3 +74,5 @@ fi
...
@@ -74,3 +74,5 @@ fi
# Show current packages
# Show current packages
$PIP_CMD
list
$PIP_CMD
list
python3
-c
"import torch; print(torch.version.cuda)"
python3
-c
"import torch; print(torch.version.cuda)"
python3
-m
flashinfer clear-cache
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment