Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
865233e2
"docs/vscode:/vscode.git/clone" did not exist on "486022e997eff0150bf93dfdd21b0dc7a397ce6f"
Unverified
Commit
865233e2
authored
Nov 23, 2024
by
Ankur Neog
Committed by
GitHub
Nov 22, 2024
Browse files
Add initial support for intel Gaudi accelerators (#2121)
parent
66d4859a
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
10 additions
and
7 deletions
+10
-7
python/pyproject.toml
python/pyproject.toml
+5
-0
python/sglang/bench_one_batch.py
python/sglang/bench_one_batch.py
+1
-4
python/sglang/srt/model_executor/model_runner.py
python/sglang/srt/model_executor/model_runner.py
+3
-2
python/sglang/srt/server_args.py
python/sglang/srt/server_args.py
+1
-1
No files found.
python/pyproject.toml
View file @
865233e2
...
...
@@ -31,6 +31,9 @@ srt_hip = ["sglang[runtime_common]", "torch", "vllm==0.6.3.dev13"]
# xpu is not enabled in public vllm and torch whl,
# need to follow https://docs.vllm.ai/en/latest/getting_started/xpu-installation.htmlinstall vllm
srt_xpu
=
["sglang[runtime_common]"]
#For Intel Gaudi(device : hpu) follow the installation guide
#https://docs.vllm.ai/en/latest/getting_started/gaudi-installation.html
srt_hpu
=
["sglang[runtime_common]"]
openai
=
[
"openai>=1.0"
,
"tiktoken"
]
anthropic
=
["anthropic>=0.20.0"]
...
...
@@ -46,9 +49,11 @@ test = [
all
=
["sglang[srt]
", "
sglang
[openai]
", "
sglang
[anthropic]
", "
sglang
[litellm]"]
all_hip
=
["sglang[srt_hip]
", "
sglang
[openai]
", "
sglang
[anthropic]
", "
sglang
[litellm]"]
all_xpu
=
["sglang[srt_xpu]
", "
sglang
[openai]
", "
sglang
[anthropic]
", "
sglang
[litellm]"]
all_hpu
=
["sglang[srt_hpu]
", "
sglang
[openai]
", "
sglang
[anthropic]
", "
sglang
[litellm]"]
dev
=
["sglang[all]
", "
sglang
[test]"]
dev_hip
=
["sglang[all_hip]
", "
sglang
[test]"]
dev_xpu
=
["sglang[all_xpu]
", "
sglang
[test]"]
dev_hpu
=
["sglang[all_hpu]
", "
sglang
[test]"]
[project.urls]
"Homepage"
=
"https://github.com/sgl-project/sglang"
...
...
python/sglang/bench_one_batch.py
View file @
865233e2
...
...
@@ -278,10 +278,7 @@ def correctness_test(
def
synchronize
(
device
):
if
device
==
"cuda"
:
torch
.
cuda
.
synchronize
()
elif
device
==
"xpu"
:
torch
.
xpu
.
synchronize
()
torch
.
get_device_module
(
device
).
synchronize
()
def
latency_test_run_once
(
...
...
python/sglang/srt/model_executor/model_runner.py
View file @
865233e2
...
...
@@ -176,14 +176,15 @@ class ModelRunner:
def
init_torch_distributed
(
self
):
logger
.
info
(
"Init torch distributed begin."
)
# Init torch distributed
torch
.
get_device_module
(
self
.
device
).
set_device
(
self
.
gpu_id
)
if
self
.
device
==
"cuda"
:
torch
.
cuda
.
set_device
(
self
.
gpu_id
)
backend
=
"nccl"
# ToDO(liangan1):Just use gloo to bypass the initilization fail
# Need to use xccl for xpu backend in the future
elif
self
.
device
==
"xpu"
:
torch
.
xpu
.
set_device
(
self
.
gpu_id
)
backend
=
"gloo"
elif
self
.
device
==
"hpu"
:
backend
=
"hccl"
if
not
self
.
server_args
.
enable_p2p_check
:
monkey_patch_vllm_p2p_access_check
(
self
.
gpu_id
)
...
...
python/sglang/srt/server_args.py
View file @
865233e2
...
...
@@ -306,7 +306,7 @@ class ServerArgs:
"--device"
,
type
=
str
,
default
=
"cuda"
,
choices
=
[
"cuda"
,
"xpu"
],
choices
=
[
"cuda"
,
"xpu"
,
"hpu"
],
help
=
"The device type."
,
)
parser
.
add_argument
(
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment