setup_rocm.py 3.08 KB
Newer Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
# Copyright 2025 SGLang Team. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================

16
import os
17
import platform
18
19
20
import sys
from pathlib import Path

21
import torch
22
23
24
25
from setuptools import find_packages, setup
from torch.utils.cpp_extension import BuildExtension, CUDAExtension

root = Path(__file__).parent.resolve()
26
arch = platform.machine().lower()
27
28
29
30
31
32
33
34
35


def _get_version():
    with open(root / "pyproject.toml") as f:
        for line in f:
            if line.startswith("version"):
                return line.split("=")[1].strip().strip('"')


36
operator_namespace = "sgl_kernel"
37
include_dirs = [
38
39
    root / "include",
    root / "csrc",
40
41
42
]

sources = [
43
    "csrc/allreduce/custom_all_reduce.hip",
44
    "csrc/allreduce/quick_all_reduce.cu",
45
    "csrc/moe/moe_align_kernel.cu",
46
    "csrc/moe/moe_topk_softmax_kernels.cu",
47
    "csrc/torch_extension_rocm.cc",
48
    "csrc/speculative/eagle_utils.cu",
49
50
51
52
]

cxx_flags = ["-O3"]
libraries = ["hiprtc", "amdhip64", "c10", "torch", "torch_python"]
53
extra_link_args = ["-Wl,-rpath,$ORIGIN/../../torch/lib", f"-L/usr/lib/{arch}-linux-gnu"]
54

55
56
57
58
59
60
61
62
63
64
65
default_target = "gfx942"
amdgpu_target = os.environ.get("AMDGPU_TARGET", default_target)

if torch.cuda.is_available():
    try:
        amdgpu_target = torch.cuda.get_device_properties(0).gcnArchName.split(":")[0]
    except Exception as e:
        print(f"Warning: Failed to detect GPU properties: {e}")
else:
    print(f"Warning: torch.cuda not available. Using default target: {amdgpu_target}")

66
67
68
69
70
71
if amdgpu_target not in ["gfx942", "gfx950"]:
    print(
        f"Warning: Unsupported GPU architecture detected '{amdgpu_target}'. Expected 'gfx942' or 'gfx950'."
    )
    sys.exit(1)

72
73
74
75
76
77
78
79
hipcc_flags = [
    "-DNDEBUG",
    f"-DOPERATOR_NAMESPACE={operator_namespace}",
    "-O3",
    "-Xcompiler",
    "-fPIC",
    "-std=c++17",
    "-D__HIP_PLATFORM_AMD__=1",
80
    f"--amdgpu-target={amdgpu_target}",
81
82
83
84
    "-DENABLE_BF16",
    "-DENABLE_FP8",
]

85
86
87
88
89
90
91
92
93
94
95
ext_modules = [
    CUDAExtension(
        name="sgl_kernel.common_ops",
        sources=sources,
        include_dirs=include_dirs,
        extra_compile_args={
            "nvcc": hipcc_flags,
            "cxx": cxx_flags,
        },
        libraries=libraries,
        extra_link_args=extra_link_args,
96
        py_limited_api=False,
97
98
99
    ),
]

100
101
102
setup(
    name="sgl-kernel",
    version=_get_version(),
103
    packages=find_packages(where="python"),
104
105
    package_dir={"": "python"},
    ext_modules=ext_modules,
106
    cmdclass={"build_ext": BuildExtension.with_options(use_ninja=True)},
107
108
    options={"bdist_wheel": {"py_limited_api": "cp39"}},
)