Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
1c63e797
Unverified
Commit
1c63e797
authored
Mar 31, 2025
by
Yineng Zhang
Committed by
GitHub
Mar 31, 2025
Browse files
use fa3 in sgl-kernel (#4954)
parent
ee47a6c1
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
3 additions
and
3 deletions
+3
-3
python/pyproject.toml
python/pyproject.toml
+1
-1
python/sglang/srt/layers/attention/flashattention_backend.py
python/sglang/srt/layers/attention/flashattention_backend.py
+1
-1
scripts/ci_install_dependency.sh
scripts/ci_install_dependency.sh
+1
-1
No files found.
python/pyproject.toml
View file @
1c63e797
...
@@ -47,7 +47,7 @@ runtime_common = [
...
@@ -47,7 +47,7 @@ runtime_common = [
srt
=
[
srt
=
[
"sglang[runtime_common]"
,
"sglang[runtime_common]"
,
"sgl-kernel==0.0.
5.post4
"
,
"sgl-kernel==0.0.
6
"
,
"flashinfer_python==0.2.3"
,
"flashinfer_python==0.2.3"
,
"torch==2.5.1"
,
"torch==2.5.1"
,
"cuda-python"
,
"cuda-python"
,
...
...
python/sglang/srt/layers/attention/flashattention_backend.py
View file @
1c63e797
...
@@ -22,7 +22,7 @@ if TYPE_CHECKING:
...
@@ -22,7 +22,7 @@ if TYPE_CHECKING:
from
sglang.srt.layers.radix_attention
import
RadixAttention
from
sglang.srt.layers.radix_attention
import
RadixAttention
from
sglang.srt.model_executor.model_runner
import
ModelRunner
from
sglang.srt.model_executor.model_runner
import
ModelRunner
from
flash_attn
_interface
import
flash_attn_with_kvcache
from
sgl_kernel.
flash_attn
import
flash_attn_with_kvcache
@
dataclass
@
dataclass
...
...
scripts/ci_install_dependency.sh
View file @
1c63e797
...
@@ -15,7 +15,7 @@ pip install -e "python[all]" --find-links https://flashinfer.ai/whl/cu124/torch2
...
@@ -15,7 +15,7 @@ pip install -e "python[all]" --find-links https://flashinfer.ai/whl/cu124/torch2
rm
-rf
/root/.cache/flashinfer
rm
-rf
/root/.cache/flashinfer
# Force reinstall flashinfer and torch_memory_saver
# Force reinstall flashinfer and torch_memory_saver
pip
install
flashinfer_python
==
0.2.3
--find-links
${
FLASHINFER_REPO
}
--force-reinstall
--no-deps
pip
install
flashinfer_python
==
0.2.3
--find-links
${
FLASHINFER_REPO
}
--force-reinstall
--no-deps
pip
install
sgl-kernel
==
0.0.
5.post4
--force-reinstall
pip
install
sgl-kernel
==
0.0.
6
--force-reinstall
pip
install
torch_memory_saver
pip
install
torch_memory_saver
pip
install
transformers
==
4.50.0 sentence_transformers
accelerate
==
1.4.0 peft pandas datasets timm torchaudio
pip
install
transformers
==
4.50.0 sentence_transformers
accelerate
==
1.4.0 peft pandas datasets timm torchaudio
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment