Unverified Commit 0b98ba15 authored by Woosuk Kwon's avatar Woosuk Kwon Committed by GitHub
Browse files

Change the name to vLLM (#150)

parent e5464ee4
from typing import Dict, List, Optional from typing import Dict, List, Optional
from cacheflow.sequence import SequenceGroup, SequenceStatus from vllm.sequence import SequenceGroup, SequenceStatus
class CompletionOutput: class CompletionOutput:
......
...@@ -2,8 +2,8 @@ import copy ...@@ -2,8 +2,8 @@ import copy
import enum import enum
from typing import Dict, List, Optional, Union from typing import Dict, List, Optional, Union
from cacheflow.block import LogicalTokenBlock from vllm.block import LogicalTokenBlock
from cacheflow.sampling_params import SamplingParams from vllm.sampling_params import SamplingParams
class SequenceStatus(enum.Enum): class SequenceStatus(enum.Enum):
......
...@@ -3,8 +3,8 @@ from typing import Dict, List, Tuple ...@@ -3,8 +3,8 @@ from typing import Dict, List, Tuple
import torch import torch
from cacheflow import cache_ops from vllm import cache_ops
from cacheflow.config import CacheConfig, ModelConfig, ParallelConfig from vllm.config import CacheConfig, ModelConfig, ParallelConfig
KVCache = Tuple[torch.Tensor, torch.Tensor] KVCache = Tuple[torch.Tensor, torch.Tensor]
......
...@@ -3,16 +3,15 @@ from typing import Dict, List, Tuple ...@@ -3,16 +3,15 @@ from typing import Dict, List, Tuple
import torch import torch
from cacheflow.config import (CacheConfig, ModelConfig, ParallelConfig, from vllm.config import (CacheConfig, ModelConfig, ParallelConfig,
SchedulerConfig) SchedulerConfig)
from cacheflow.model_executor import get_model, InputMetadata, set_random_seed from vllm.model_executor import get_model, InputMetadata, set_random_seed
from cacheflow.model_executor.parallel_utils.parallel_state import ( from vllm.model_executor.parallel_utils.parallel_state import (
initialize_model_parallel, initialize_all_reduce_launcher) initialize_model_parallel, initialize_all_reduce_launcher)
from cacheflow.sampling_params import SamplingParams from vllm.sampling_params import SamplingParams
from cacheflow.sequence import (SequenceData, SequenceGroupMetadata, from vllm.sequence import SequenceData, SequenceGroupMetadata, SequenceOutputs
SequenceOutputs) from vllm.worker.cache_engine import CacheEngine
from cacheflow.worker.cache_engine import CacheEngine from vllm.utils import get_gpu_memory
from cacheflow.utils import get_gpu_memory
class Worker: class Worker:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment