pyproject.toml 3.53 KB
Newer Older
Lianmin Zheng's avatar
Lianmin Zheng committed
1
2
3
4
5
6
[build-system]
requires = ["setuptools>=61.0", "wheel"]
build-backend = "setuptools.build_meta"

[project]
name = "sglang"
7
version = "0.4.3.post2"
8
description = "SGLang is yet another fast serving framework for large language models and vision language models."
Lianmin Zheng's avatar
Lianmin Zheng committed
9
10
readme = "README.md"
requires-python = ">=3.8"
11
license = { file = "LICENSE" }
Lianmin Zheng's avatar
Lianmin Zheng committed
12
13
14
15
classifiers = [
    "Programming Language :: Python :: 3",
    "License :: OSI Approved :: Apache Software License",
]
16
dependencies = ["requests", "tqdm", "numpy", "IPython", "setproctitle"]
Lianmin Zheng's avatar
Lianmin Zheng committed
17
18

[project.optional-dependencies]
19
runtime_common = [
Lianmin Zheng's avatar
Lianmin Zheng committed
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
    "aiohttp",
    "decord",
    "fastapi",
    "hf_transfer",
    "huggingface_hub",
    "interegular",
    "modelscope",
    "orjson",
    "packaging",
    "pillow",
    "prometheus-client>=0.20.0",
    "psutil",
    "pydantic",
    "python-multipart",
    "pyzmq>=25.1.2",
    "torchao>=0.7.0",
    "uvicorn",
    "uvloop",
    "xgrammar==0.1.10",
    "ninja",
    "transformers==4.48.3",
41
42
]
srt = [
Lianmin Zheng's avatar
Lianmin Zheng committed
43
44
    "sglang[runtime_common]",
    "sgl-kernel>=0.0.3.post6",
45
    "flashinfer_python>=0.2.1.post2",
Lianmin Zheng's avatar
Lianmin Zheng committed
46
47
48
    "torch==2.5.1",
    "vllm>=0.6.4.post1,<=0.7.2",
    "cuda-python",
49
    "outlines>=0.0.44,<=0.1.11",
50
]
51

52
53
# HIP (Heterogeneous-computing Interface for Portability) for AMD
# => base docker rocm/vllm-dev:20241022, not from public vllm whl
Lianmin Zheng's avatar
Lianmin Zheng committed
54
55
srt_hip = ["sglang[runtime_common]", "sgl-kernel>=0.0.3.post1", "torch", "vllm==0.6.7.dev2", "outlines==0.1.11"]

56
57
# xpu is not enabled in public vllm and torch whl,
# need to follow https://docs.vllm.ai/en/latest/getting_started/xpu-installation.htmlinstall vllm
Lianmin Zheng's avatar
Lianmin Zheng committed
58
59
60
61
62
63
srt_xpu = ["sglang[runtime_common]", "outlines>=0.0.44,<=0.1.11"]

# For Intel Gaudi(device : hpu) follow the installation guide
# https://docs.vllm.ai/en/latest/getting_started/gaudi-installation.html
srt_hpu = ["sglang[runtime_common]", "outlines>=0.0.44,<=0.1.11"]

64
65
66
# CPU: currently, there are no pre-built vllm wheels for CPU.
# To install vllm for CPU, please follow the instruction here:
# https://docs.vllm.ai/en/latest/getting_started/installation/cpu/index.html
Lianmin Zheng's avatar
Lianmin Zheng committed
67
srt_cpu = ["sglang[runtime_common]", "outlines>=0.0.44,<=0.1.11", "torch"]
68

69
70
openai = ["openai>=1.0", "tiktoken"]
anthropic = ["anthropic>=0.20.0"]
胡译文's avatar
胡译文 committed
71
litellm = ["litellm>=1.0.0"]
72
torch_memory_saver = ["torch_memory_saver"]
73
74
75
76
77
78
79
80
test = [
    "jsonlines",
    "matplotlib",
    "pandas",
    "sentence_transformers",
    "accelerate",
    "peft",
]
胡译文's avatar
胡译文 committed
81
all = ["sglang[srt]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
82
all_hip = ["sglang[srt_hip]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
83
all_xpu = ["sglang[srt_xpu]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
84
all_hpu = ["sglang[srt_hpu]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
85
all_cpu = ["sglang[srt_cpu]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
86

87
dev = ["sglang[all]", "sglang[test]"]
88
dev_hip = ["sglang[all_hip]", "sglang[test]"]
89
dev_xpu = ["sglang[all_xpu]", "sglang[test]"]
90
dev_hpu = ["sglang[all_hpu]", "sglang[test]"]
91
dev_cpu = ["sglang[all_cpu]", "sglang[test]"]
Lianmin Zheng's avatar
Lianmin Zheng committed
92

Lianmin Zheng's avatar
Lianmin Zheng committed
93
94
95
96
[project.urls]
"Homepage" = "https://github.com/sgl-project/sglang"
"Bug Tracker" = "https://github.com/sgl-project/sglang/issues"

yudian0504's avatar
yudian0504 committed
97
[tool.setuptools.package-data]
98
"sglang" = ["srt/layers/moe/fused_moe_triton/configs/*.json", "srt/layers/quantization/configs/*.json"]
yudian0504's avatar
yudian0504 committed
99

Lianmin Zheng's avatar
Lianmin Zheng committed
100
[tool.setuptools.packages.find]
101
102
103
104
105
106
107
108
109
exclude = [
    "assets*",
    "benchmark*",
    "docs*",
    "dist*",
    "playground*",
    "scripts*",
    "tests*",
]
Lianmin Zheng's avatar
Lianmin Zheng committed
110
111

[tool.wheel]
112
113
114
115
116
117
118
119
120
exclude = [
    "assets*",
    "benchmark*",
    "docs*",
    "dist*",
    "playground*",
    "scripts*",
    "tests*",
]