Unverified Commit b149b393 authored by Brayden Zhong's avatar Brayden Zhong Committed by GitHub
Browse files

[CI] Remove unused imports with Ruff to pre-commit config, only to...

[CI] Remove unused imports with Ruff to pre-commit config, only to benchmarks/docs/examples folder (#3969)
parent 31dfff7d
...@@ -22,6 +22,13 @@ repos: ...@@ -22,6 +22,13 @@ repos:
rev: 5.13.2 rev: 5.13.2
hooks: hooks:
- id: isort - id: isort
- repo: https://github.com/astral-sh/ruff-pre-commit
rev: v0.11.2
hooks:
- id: ruff
args: [--select=F401, --fixable=F401]
files: ^(benchmark/|docs/|examples/)
exclude: \.ipynb$
- repo: https://github.com/psf/black - repo: https://github.com/psf/black
rev: 24.10.0 rev: 24.10.0
hooks: hooks:
......
...@@ -23,7 +23,7 @@ import warnings ...@@ -23,7 +23,7 @@ import warnings
from argparse import ArgumentParser from argparse import ArgumentParser
from dataclasses import dataclass, field from dataclasses import dataclass, field
from datetime import datetime from datetime import datetime
from typing import Any, AsyncGenerator, Dict, List, Optional, Tuple, Union from typing import Any, AsyncGenerator, Dict, List, Optional, Tuple
import aiohttp import aiohttp
import numpy as np import numpy as np
......
...@@ -4,8 +4,6 @@ import math ...@@ -4,8 +4,6 @@ import math
import cudnn import cudnn
import torch import torch
import torch.utils.benchmark as benchmark import torch.utils.benchmark as benchmark
import triton
import triton.language as tl
from flashinfer import BatchDecodeWithPagedKVCacheWrapper from flashinfer import BatchDecodeWithPagedKVCacheWrapper
from sglang.srt.layers.attention.triton_ops.decode_attention import decode_attention_fwd from sglang.srt.layers.attention.triton_ops.decode_attention import decode_attention_fwd
......
import itertools import itertools
import math import math
import os
from typing import Optional, Tuple from typing import Optional, Tuple
import torch import torch
......
...@@ -3,7 +3,6 @@ from typing import Optional, Tuple, Union ...@@ -3,7 +3,6 @@ from typing import Optional, Tuple, Union
import torch import torch
import triton import triton
import triton.language as tl
from flashinfer.norm import fused_add_rmsnorm, rmsnorm from flashinfer.norm import fused_add_rmsnorm, rmsnorm
from torch import nn from torch import nn
from vllm import _custom_ops as vllm_ops from vllm import _custom_ops as vllm_ops
......
import itertools import itertools
import os import os
from typing import List
import numpy as np
import pytest
import torch import torch
import triton import triton
import triton.language as tl import triton.language as tl
......
...@@ -15,42 +15,28 @@ ...@@ -15,42 +15,28 @@
import argparse import argparse
import asyncio import asyncio
import json import json
import os
import random import random
import resource import resource
import sys import sys
import time import time
import traceback import traceback
import warnings
from argparse import ArgumentParser from argparse import ArgumentParser
from dataclasses import dataclass, field
from datetime import datetime from datetime import datetime
from typing import Any, AsyncGenerator, Dict, List, Optional, Tuple, Union from typing import Any, Dict, List, Optional, Tuple
import aiohttp import aiohttp
import numpy as np import numpy as np
import requests
from launch_server import LORA_PATH, NUM_LORAS from launch_server import LORA_PATH, NUM_LORAS
from tqdm.asyncio import tqdm from tqdm.asyncio import tqdm
from transformers import ( from transformers import PreTrainedTokenizerBase
AutoTokenizer,
PreTrainedTokenizer,
PreTrainedTokenizerBase,
PreTrainedTokenizerFast,
)
from sglang.bench_serving import ( from sglang.bench_serving import (
AIOHTTP_TIMEOUT, AIOHTTP_TIMEOUT,
SHAREGPT_URL,
BenchmarkMetrics,
RequestFuncInput, RequestFuncInput,
RequestFuncOutput, RequestFuncOutput,
calculate_metrics, calculate_metrics,
check_chat_template,
get_model,
get_request, get_request,
get_tokenizer, get_tokenizer,
parse_request_rate_range,
remove_prefix, remove_prefix,
sample_random_requests, sample_random_requests,
) )
......
...@@ -6,7 +6,6 @@ import time ...@@ -6,7 +6,6 @@ import time
import numpy as np import numpy as np
import pandas as pd import pandas as pd
import tiktoken import tiktoken
from tqdm import tqdm
from sglang.test.test_utils import ( from sglang.test.test_utils import (
add_common_sglang_args_and_parse, add_common_sglang_args_and_parse,
......
import argparse import argparse
import PIL.Image
import torch import torch
from data_utils import save_json from data_utils import save_json
from eval_utils import ( from eval_utils import (
......
...@@ -5,7 +5,6 @@ import os ...@@ -5,7 +5,6 @@ import os
import re import re
import yaml import yaml
from datasets import concatenate_datasets, load_dataset
DOMAIN_CAT2SUB_CAT = { DOMAIN_CAT2SUB_CAT = {
"Art and Design": ["Art", "Art_Theory", "Design", "Music"], "Art and Design": ["Art", "Art_Theory", "Design", "Music"],
......
import itertools
import json import json
import os
import random import random
import string
import threading
import time import time
from argparse import ArgumentParser from argparse import ArgumentParser
from pathlib import Path from pathlib import Path
from typing import Union
from tqdm import tqdm from tqdm import tqdm
......
import os
import weakref import weakref
import nest_asyncio import nest_asyncio
......
import os
import weakref import weakref
from sglang.utils import execute_shell_command, reserve_port from sglang.utils import execute_shell_command, reserve_port
......
...@@ -4,8 +4,6 @@ export OPENAI_API_KEY=sk-****** ...@@ -4,8 +4,6 @@ export OPENAI_API_KEY=sk-******
python3 openai_example_chat.py python3 openai_example_chat.py
""" """
import json
import sglang as sgl import sglang as sgl
......
# NOTE: Currently this can only be run through HTTP requests. # NOTE: Currently this can only be run through HTTP requests.
import json
from concurrent.futures import ThreadPoolExecutor from concurrent.futures import ThreadPoolExecutor
from json_decode import character_regex from json_decode import character_regex
......
...@@ -5,11 +5,6 @@ python offline_batch_inference_vlm.py --model-path Qwen/Qwen2-VL-7B-Instruct --c ...@@ -5,11 +5,6 @@ python offline_batch_inference_vlm.py --model-path Qwen/Qwen2-VL-7B-Instruct --c
import argparse import argparse
import dataclasses import dataclasses
import io
import os
import requests
from PIL import Image
import sglang as sgl import sglang as sgl
from sglang.srt.conversation import chat_templates from sglang.srt.conversation import chat_templates
......
...@@ -12,7 +12,7 @@ import requests ...@@ -12,7 +12,7 @@ import requests
import torch import torch
from sglang.test.test_utils import is_in_ci from sglang.test.test_utils import is_in_ci
from sglang.utils import print_highlight, terminate_process, wait_for_server from sglang.utils import terminate_process, wait_for_server
if is_in_ci(): if is_in_ci():
from docs.backend.patch import launch_server_cmd from docs.backend.patch import launch_server_cmd
......
...@@ -11,7 +11,6 @@ you should create the input.jsonl file with the following content: ...@@ -11,7 +11,6 @@ you should create the input.jsonl file with the following content:
""" """
import json import json
import os
import time import time
import openai import openai
......
...@@ -5,7 +5,6 @@ python openai_chat.py ...@@ -5,7 +5,6 @@ python openai_chat.py
""" """
import openai import openai
from openai import OpenAI
client = openai.Client(base_url="http://127.0.0.1:30000/v1", api_key="EMPTY") client = openai.Client(base_url="http://127.0.0.1:30000/v1", api_key="EMPTY")
......
...@@ -9,7 +9,7 @@ import requests ...@@ -9,7 +9,7 @@ import requests
from sglang.srt.hf_transformers_utils import get_tokenizer from sglang.srt.hf_transformers_utils import get_tokenizer
from sglang.test.test_utils import is_in_ci from sglang.test.test_utils import is_in_ci
from sglang.utils import print_highlight, terminate_process, wait_for_server from sglang.utils import terminate_process, wait_for_server
if is_in_ci(): if is_in_ci():
from docs.backend.patch import launch_server_cmd from docs.backend.patch import launch_server_cmd
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment