Unverified Commit 150d7020 authored by Liangsheng Yin's avatar Liangsheng Yin Committed by GitHub
Browse files

Revert removing the unused imports (#385)

parent 9acc6e35
...@@ -3,7 +3,11 @@ ...@@ -3,7 +3,11 @@
import re import re
from typing import Callable, List, Optional, Union from typing import Callable, List, Optional, Union
from sglang.backend.anthropic import Anthropic
from sglang.backend.base_backend import BaseBackend from sglang.backend.base_backend import BaseBackend
from sglang.backend.openai import OpenAI
from sglang.backend.runtime_endpoint import RuntimeEndpoint
from sglang.backend.vertexai import VertexAI
from sglang.global_config import global_config from sglang.global_config import global_config
from sglang.lang.ir import ( from sglang.lang.ir import (
SglExpr, SglExpr,
......
from typing import List, Optional, Union
import numpy as np
from sglang.backend.base_backend import BaseBackend from sglang.backend.base_backend import BaseBackend
from sglang.lang.chat_template import get_chat_template from sglang.lang.chat_template import get_chat_template
from sglang.lang.interpreter import StreamExecutor from sglang.lang.interpreter import StreamExecutor
......
from typing import List, Optional, Union from typing import Callable, List, Optional, Union
from sglang.lang.chat_template import get_chat_template from sglang.lang.chat_template import get_chat_template
from sglang.lang.interpreter import StreamExecutor from sglang.lang.interpreter import StreamExecutor
......
import logging import logging
import time import time
from typing import List, Optional from typing import Callable, List, Optional, Union
import numpy as np import numpy as np
......
import json import json
from typing import List, Optional from typing import Callable, List, Optional, Union
import numpy as np import numpy as np
import requests
from sglang.backend.base_backend import BaseBackend from sglang.backend.base_backend import BaseBackend
from sglang.global_config import global_config from sglang.global_config import global_config
from sglang.lang.chat_template import get_chat_template_by_model_path from sglang.lang.chat_template import get_chat_template_by_model_path
from sglang.lang.interpreter import StreamExecutor from sglang.lang.interpreter import StreamExecutor
from sglang.lang.ir import SglSamplingParams from sglang.lang.ir import SglArgument, SglSamplingParams
from sglang.utils import find_printable_text, http_request from sglang.utils import encode_image_base64, find_printable_text, http_request
class RuntimeEndpoint(BaseBackend): class RuntimeEndpoint(BaseBackend):
......
import os import os
import warnings import warnings
from typing import List, Optional, Union
import numpy as np
from sglang.backend.base_backend import BaseBackend from sglang.backend.base_backend import BaseBackend
from sglang.lang.chat_template import get_chat_template from sglang.lang.chat_template import get_chat_template
......
from dataclasses import dataclass from dataclasses import dataclass, field
from enum import Enum, auto from enum import Enum, auto
from typing import Callable, Dict, List, Tuple from typing import Callable, Dict, List, Optional, Tuple
class ChatTemplateStyle(Enum): class ChatTemplateStyle(Enum):
......
...@@ -5,7 +5,13 @@ from typing import List, Union ...@@ -5,7 +5,13 @@ from typing import List, Union
from sglang.global_config import global_config from sglang.global_config import global_config
from sglang.lang.interpreter import ProgramState, StreamExecutor, pin_program from sglang.lang.interpreter import ProgramState, StreamExecutor, pin_program
from sglang.lang.ir import SglArgument, SglExpr, SglSamplingParams, SglVariable from sglang.lang.ir import (
SglArgument,
SglConstantText,
SglExpr,
SglSamplingParams,
SglVariable,
)
def compile_func(function, backend): def compile_func(function, backend):
......
...@@ -7,7 +7,7 @@ import threading ...@@ -7,7 +7,7 @@ import threading
import uuid import uuid
from concurrent.futures import ThreadPoolExecutor from concurrent.futures import ThreadPoolExecutor
from contextlib import contextmanager from contextlib import contextmanager
from typing import Any, Callable, Dict, List, Optional from typing import Any, Callable, Dict, List, Optional, Union
import tqdm import tqdm
...@@ -18,6 +18,7 @@ from sglang.lang.ir import ( ...@@ -18,6 +18,7 @@ from sglang.lang.ir import (
SglConstantText, SglConstantText,
SglExpr, SglExpr,
SglExprList, SglExprList,
SglFunction,
SglGen, SglGen,
SglImage, SglImage,
SglRoleBegin, SglRoleBegin,
......
...@@ -472,4 +472,4 @@ class SglCommitLazy(SglExpr): ...@@ -472,4 +472,4 @@ class SglCommitLazy(SglExpr):
super().__init__() super().__init__()
def __repr__(self): def __repr__(self):
return "CommitLazy()" return f"CommitLazy()"
"""Tracing a program.""" """Tracing a program."""
import uuid import uuid
from typing import Any, Dict, List, Optional from typing import Any, Callable, Dict, List, Optional, Union
from sglang.backend.base_backend import BaseBackend from sglang.backend.base_backend import BaseBackend
from sglang.global_config import global_config
from sglang.lang.interpreter import ProgramState, ProgramStateGroup from sglang.lang.interpreter import ProgramState, ProgramStateGroup
from sglang.lang.ir import ( from sglang.lang.ir import (
SglArgument, SglArgument,
SglCommitLazy,
SglConcateAndAppend,
SglConstantText, SglConstantText,
SglExpr, SglExpr,
SglExprList, SglExprList,
SglFork, SglFork,
SglFunction,
SglGen, SglGen,
SglGetForkItem, SglGetForkItem,
SglRoleBegin, SglRoleBegin,
......
...@@ -3,7 +3,7 @@ ...@@ -3,7 +3,7 @@
import json import json
import os import os
import warnings import warnings
from typing import Optional, Union from typing import List, Optional, Tuple, Union
from huggingface_hub import snapshot_download from huggingface_hub import snapshot_download
from transformers import ( from transformers import (
......
...@@ -84,7 +84,7 @@ def start_detokenizer_process( ...@@ -84,7 +84,7 @@ def start_detokenizer_process(
): ):
try: try:
manager = DetokenizerManager(server_args, port_args) manager = DetokenizerManager(server_args, port_args)
except Exception: except Exception as e:
pipe_writer.send(get_exception_traceback()) pipe_writer.send(get_exception_traceback())
raise raise
pipe_writer.send("init ok") pipe_writer.send("init ok")
......
import heapq import heapq
import time import time
from collections import defaultdict from collections import defaultdict
from dataclasses import dataclass
from typing import Tuple
import torch import torch
......
...@@ -20,7 +20,7 @@ ...@@ -20,7 +20,7 @@
# This file is based on the LLama model definition file in transformers # This file is based on the LLama model definition file in transformers
"""PyTorch Cohere model.""" """PyTorch Cohere model."""
from typing import Optional, Tuple from typing import List, Optional, Tuple
import torch import torch
import torch.utils.checkpoint import torch.utils.checkpoint
......
# Adapted from # Adapted from
# https://github.com/vllm-project/vllm/blob/671af2b1c0b3ed6d856d37c21a561cc429a10701/vllm/model_executor/models/llama.py#L1 # https://github.com/vllm-project/vllm/blob/671af2b1c0b3ed6d856d37c21a561cc429a10701/vllm/model_executor/models/llama.py#L1
"""Inference-only LLaMA model compatible with HuggingFace weights.""" """Inference-only LLaMA model compatible with HuggingFace weights."""
from typing import Any, Dict, Optional, Tuple from typing import Any, Dict, List, Optional, Tuple
import torch import torch
from torch import nn from torch import nn
......
...@@ -5,7 +5,7 @@ from typing import List, Optional ...@@ -5,7 +5,7 @@ from typing import List, Optional
import numpy as np import numpy as np
import torch import torch
from torch import nn from torch import nn
from transformers import CLIPVisionModel, LlavaConfig from transformers import CLIPVisionModel, LlamaConfig, LlavaConfig
from transformers.models.llava.modeling_llava import LlavaMultiModalProjector from transformers.models.llava.modeling_llava import LlavaMultiModalProjector
from vllm.model_executor.layers.linear import LinearMethodBase from vllm.model_executor.layers.linear import LinearMethodBase
from vllm.model_executor.weight_utils import ( from vllm.model_executor.weight_utils import (
......
# Adapted from # Adapted from
# https://github.com/vllm-project/vllm/blob/d0215a58e78572d91dadafe9d832a2db89b09a13/vllm/model_executor/models/mixtral.py#L1 # https://github.com/vllm-project/vllm/blob/d0215a58e78572d91dadafe9d832a2db89b09a13/vllm/model_executor/models/mixtral.py#L1
"""Inference-only Mixtral model.""" """Inference-only Mixtral model."""
from typing import Optional from typing import List, Optional, Tuple
import numpy as np import numpy as np
import torch import torch
......
from typing import Any, Dict, Optional from typing import Any, Dict, List, Optional, Tuple
import torch import torch
from torch import nn from torch import nn
......
# Adapted from llama2.py # Adapted from llama2.py
# Modify details for the adaptation of Qwen2 model. # Modify details for the adaptation of Qwen2 model.
"""Inference-only Qwen2 model compatible with HuggingFace weights.""" """Inference-only Qwen2 model compatible with HuggingFace weights."""
from typing import Any, Dict, Optional, Tuple from typing import Any, Dict, List, Optional, Tuple
import torch import torch
from torch import nn from torch import nn
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment