Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
norm
vllm
Commits
0b98ba15
Unverified
Commit
0b98ba15
authored
Jun 17, 2023
by
Woosuk Kwon
Committed by
GitHub
Jun 17, 2023
Browse files
Change the name to vLLM (#150)
parent
e5464ee4
Changes
88
Hide whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
13 additions
and
14 deletions
+13
-14
vllm/model_executor/weight_utils.py
vllm/model_executor/weight_utils.py
+0
-0
vllm/outputs.py
vllm/outputs.py
+1
-1
vllm/sampling_params.py
vllm/sampling_params.py
+0
-0
vllm/sequence.py
vllm/sequence.py
+2
-2
vllm/utils.py
vllm/utils.py
+0
-0
vllm/worker/__init__.py
vllm/worker/__init__.py
+0
-0
vllm/worker/cache_engine.py
vllm/worker/cache_engine.py
+2
-2
vllm/worker/worker.py
vllm/worker/worker.py
+8
-9
No files found.
cacheflow
/model_executor/weight_utils.py
→
vllm
/model_executor/weight_utils.py
View file @
0b98ba15
File moved
cacheflow
/outputs.py
→
vllm
/outputs.py
View file @
0b98ba15
from
typing
import
Dict
,
List
,
Optional
from
typing
import
Dict
,
List
,
Optional
from
cacheflow
.sequence
import
SequenceGroup
,
SequenceStatus
from
vllm
.sequence
import
SequenceGroup
,
SequenceStatus
class
CompletionOutput
:
class
CompletionOutput
:
...
...
cacheflow
/sampling_params.py
→
vllm
/sampling_params.py
View file @
0b98ba15
File moved
cacheflow
/sequence.py
→
vllm
/sequence.py
View file @
0b98ba15
...
@@ -2,8 +2,8 @@ import copy
...
@@ -2,8 +2,8 @@ import copy
import
enum
import
enum
from
typing
import
Dict
,
List
,
Optional
,
Union
from
typing
import
Dict
,
List
,
Optional
,
Union
from
cacheflow
.block
import
LogicalTokenBlock
from
vllm
.block
import
LogicalTokenBlock
from
cacheflow
.sampling_params
import
SamplingParams
from
vllm
.sampling_params
import
SamplingParams
class
SequenceStatus
(
enum
.
Enum
):
class
SequenceStatus
(
enum
.
Enum
):
...
...
cacheflow
/utils.py
→
vllm
/utils.py
View file @
0b98ba15
File moved
cacheflow
/worker/__init__.py
→
vllm
/worker/__init__.py
View file @
0b98ba15
File moved
cacheflow
/worker/cache_engine.py
→
vllm
/worker/cache_engine.py
View file @
0b98ba15
...
@@ -3,8 +3,8 @@ from typing import Dict, List, Tuple
...
@@ -3,8 +3,8 @@ from typing import Dict, List, Tuple
import
torch
import
torch
from
cacheflow
import
cache_ops
from
vllm
import
cache_ops
from
cacheflow
.config
import
CacheConfig
,
ModelConfig
,
ParallelConfig
from
vllm
.config
import
CacheConfig
,
ModelConfig
,
ParallelConfig
KVCache
=
Tuple
[
torch
.
Tensor
,
torch
.
Tensor
]
KVCache
=
Tuple
[
torch
.
Tensor
,
torch
.
Tensor
]
...
...
cacheflow
/worker/worker.py
→
vllm
/worker/worker.py
View file @
0b98ba15
...
@@ -3,16 +3,15 @@ from typing import Dict, List, Tuple
...
@@ -3,16 +3,15 @@ from typing import Dict, List, Tuple
import
torch
import
torch
from
cacheflow
.config
import
(
CacheConfig
,
ModelConfig
,
ParallelConfig
,
from
vllm
.config
import
(
CacheConfig
,
ModelConfig
,
ParallelConfig
,
SchedulerConfig
)
SchedulerConfig
)
from
cacheflow
.model_executor
import
get_model
,
InputMetadata
,
set_random_seed
from
vllm
.model_executor
import
get_model
,
InputMetadata
,
set_random_seed
from
cacheflow
.model_executor.parallel_utils.parallel_state
import
(
from
vllm
.model_executor.parallel_utils.parallel_state
import
(
initialize_model_parallel
,
initialize_all_reduce_launcher
)
initialize_model_parallel
,
initialize_all_reduce_launcher
)
from
cacheflow.sampling_params
import
SamplingParams
from
vllm.sampling_params
import
SamplingParams
from
cacheflow.sequence
import
(
SequenceData
,
SequenceGroupMetadata
,
from
vllm.sequence
import
SequenceData
,
SequenceGroupMetadata
,
SequenceOutputs
SequenceOutputs
)
from
vllm.worker.cache_engine
import
CacheEngine
from
cacheflow.worker.cache_engine
import
CacheEngine
from
vllm.utils
import
get_gpu_memory
from
cacheflow.utils
import
get_gpu_memory
class
Worker
:
class
Worker
:
...
...
Prev
1
2
3
4
5
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment