pyproject.toml 2.9 KB
Newer Older
Lianmin Zheng's avatar
Lianmin Zheng committed
1
2
3
4
5
6
[build-system]
requires = ["setuptools>=61.0", "wheel"]
build-backend = "setuptools.build_meta"

[project]
name = "sglang"
7
version = "0.4.1.post6"
8
description = "SGLang is yet another fast serving framework for large language models and vision language models."
Lianmin Zheng's avatar
Lianmin Zheng committed
9
10
readme = "README.md"
requires-python = ">=3.8"
11
license = { file = "LICENSE" }
Lianmin Zheng's avatar
Lianmin Zheng committed
12
13
14
15
classifiers = [
    "Programming Language :: Python :: 3",
    "License :: OSI Approved :: Apache Software License",
]
16
dependencies = ["requests", "tqdm", "numpy", "IPython", "setproctitle"]
Lianmin Zheng's avatar
Lianmin Zheng committed
17
18

[project.optional-dependencies]
19
20
runtime_common = [
    "aiohttp", "decord", "fastapi",
Lianmin Zheng's avatar
Lianmin Zheng committed
21
    "hf_transfer", "huggingface_hub", "interegular", "modelscope",
22
23
24
    "orjson", "outlines>=0.0.44,<0.1.0",
    "packaging", "pillow", "prometheus-client>=0.20.0",
    "psutil", "pydantic", "python-multipart",
25
    "pyzmq>=25.1.2", "torchao>=0.7.0", "uvicorn", "uvloop",
26
27
28
29
    "xgrammar>=0.1.6"
]
srt = [
    "sglang[runtime_common]", "cuda-python",
30
    "sgl-kernel>=0.0.2.post14", "torch", "vllm>=0.6.3.post1,<=0.6.4.post1",
31
    "flashinfer==0.1.6"
32
]
33

34
35
# HIP (Heterogeneous-computing Interface for Portability) for AMD
# => base docker rocm/vllm-dev:20241022, not from public vllm whl
kk's avatar
kk committed
36
srt_hip = ["sglang[runtime_common]", "torch", "vllm==0.6.3.post2.dev1"]
37
38
39
# xpu is not enabled in public vllm and torch whl,
# need to follow https://docs.vllm.ai/en/latest/getting_started/xpu-installation.htmlinstall vllm
srt_xpu = ["sglang[runtime_common]"]
40
41
#For Intel Gaudi(device : hpu) follow the installation guide
#https://docs.vllm.ai/en/latest/getting_started/gaudi-installation.html
42
srt_hpu = ["sglang[runtime_common]"]
43

44
45
openai = ["openai>=1.0", "tiktoken"]
anthropic = ["anthropic>=0.20.0"]
胡译文's avatar
胡译文 committed
46
litellm = ["litellm>=1.0.0"]
47
torch_memory_saver = ["torch_memory_saver"]
48
49
50
51
52
53
54
55
test = [
    "jsonlines",
    "matplotlib",
    "pandas",
    "sentence_transformers",
    "accelerate",
    "peft",
]
胡译文's avatar
胡译文 committed
56
all = ["sglang[srt]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
57
all_hip = ["sglang[srt_hip]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
58
all_xpu = ["sglang[srt_xpu]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
59
all_hpu = ["sglang[srt_hpu]", "sglang[openai]", "sglang[anthropic]", "sglang[litellm]"]
60

61
dev = ["sglang[all]", "sglang[test]"]
62
dev_hip = ["sglang[all_hip]", "sglang[test]"]
63
dev_xpu = ["sglang[all_xpu]", "sglang[test]"]
64
dev_hpu = ["sglang[all_hpu]", "sglang[test]"]
Lianmin Zheng's avatar
Lianmin Zheng committed
65

Lianmin Zheng's avatar
Lianmin Zheng committed
66
67
68
69
[project.urls]
"Homepage" = "https://github.com/sgl-project/sglang"
"Bug Tracker" = "https://github.com/sgl-project/sglang/issues"

yudian0504's avatar
yudian0504 committed
70
[tool.setuptools.package-data]
71
"sglang" = ["srt/layers/moe/fused_moe_triton/configs/*.json", "srt/layers/quantization/configs/*.json"]
yudian0504's avatar
yudian0504 committed
72

Lianmin Zheng's avatar
Lianmin Zheng committed
73
[tool.setuptools.packages.find]
74
75
76
77
78
79
80
81
82
exclude = [
    "assets*",
    "benchmark*",
    "docs*",
    "dist*",
    "playground*",
    "scripts*",
    "tests*",
]
Lianmin Zheng's avatar
Lianmin Zheng committed
83
84

[tool.wheel]
85
86
87
88
89
90
91
92
93
exclude = [
    "assets*",
    "benchmark*",
    "docs*",
    "dist*",
    "playground*",
    "scripts*",
    "tests*",
]