Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
norm
vllm
Commits
0b98ba15
"example/49_fpAintB_gemm/fp16int8_gemm_wmma.cpp" did not exist on "ec7c2e912e1c101ea8bad335f1f22670f448776c"
Unverified
Commit
0b98ba15
authored
Jun 17, 2023
by
Woosuk Kwon
Committed by
GitHub
Jun 17, 2023
Browse files
Change the name to vLLM (#150)
parent
e5464ee4
Changes
88
Show whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
13 additions
and
14 deletions
+13
-14
vllm/model_executor/weight_utils.py
vllm/model_executor/weight_utils.py
+0
-0
vllm/outputs.py
vllm/outputs.py
+1
-1
vllm/sampling_params.py
vllm/sampling_params.py
+0
-0
vllm/sequence.py
vllm/sequence.py
+2
-2
vllm/utils.py
vllm/utils.py
+0
-0
vllm/worker/__init__.py
vllm/worker/__init__.py
+0
-0
vllm/worker/cache_engine.py
vllm/worker/cache_engine.py
+2
-2
vllm/worker/worker.py
vllm/worker/worker.py
+8
-9
No files found.
cacheflow
/model_executor/weight_utils.py
→
vllm
/model_executor/weight_utils.py
View file @
0b98ba15
File moved
cacheflow
/outputs.py
→
vllm
/outputs.py
View file @
0b98ba15
from
typing
import
Dict
,
List
,
Optional
from
cacheflow
.sequence
import
SequenceGroup
,
SequenceStatus
from
vllm
.sequence
import
SequenceGroup
,
SequenceStatus
class
CompletionOutput
:
...
...
cacheflow
/sampling_params.py
→
vllm
/sampling_params.py
View file @
0b98ba15
File moved
cacheflow
/sequence.py
→
vllm
/sequence.py
View file @
0b98ba15
...
...
@@ -2,8 +2,8 @@ import copy
import
enum
from
typing
import
Dict
,
List
,
Optional
,
Union
from
cacheflow
.block
import
LogicalTokenBlock
from
cacheflow
.sampling_params
import
SamplingParams
from
vllm
.block
import
LogicalTokenBlock
from
vllm
.sampling_params
import
SamplingParams
class
SequenceStatus
(
enum
.
Enum
):
...
...
cacheflow
/utils.py
→
vllm
/utils.py
View file @
0b98ba15
File moved
cacheflow
/worker/__init__.py
→
vllm
/worker/__init__.py
View file @
0b98ba15
File moved
cacheflow
/worker/cache_engine.py
→
vllm
/worker/cache_engine.py
View file @
0b98ba15
...
...
@@ -3,8 +3,8 @@ from typing import Dict, List, Tuple
import
torch
from
cacheflow
import
cache_ops
from
cacheflow
.config
import
CacheConfig
,
ModelConfig
,
ParallelConfig
from
vllm
import
cache_ops
from
vllm
.config
import
CacheConfig
,
ModelConfig
,
ParallelConfig
KVCache
=
Tuple
[
torch
.
Tensor
,
torch
.
Tensor
]
...
...
cacheflow
/worker/worker.py
→
vllm
/worker/worker.py
View file @
0b98ba15
...
...
@@ -3,16 +3,15 @@ from typing import Dict, List, Tuple
import
torch
from
cacheflow
.config
import
(
CacheConfig
,
ModelConfig
,
ParallelConfig
,
from
vllm
.config
import
(
CacheConfig
,
ModelConfig
,
ParallelConfig
,
SchedulerConfig
)
from
cacheflow
.model_executor
import
get_model
,
InputMetadata
,
set_random_seed
from
cacheflow
.model_executor.parallel_utils.parallel_state
import
(
from
vllm
.model_executor
import
get_model
,
InputMetadata
,
set_random_seed
from
vllm
.model_executor.parallel_utils.parallel_state
import
(
initialize_model_parallel
,
initialize_all_reduce_launcher
)
from
cacheflow.sampling_params
import
SamplingParams
from
cacheflow.sequence
import
(
SequenceData
,
SequenceGroupMetadata
,
SequenceOutputs
)
from
cacheflow.worker.cache_engine
import
CacheEngine
from
cacheflow.utils
import
get_gpu_memory
from
vllm.sampling_params
import
SamplingParams
from
vllm.sequence
import
SequenceData
,
SequenceGroupMetadata
,
SequenceOutputs
from
vllm.worker.cache_engine
import
CacheEngine
from
vllm.utils
import
get_gpu_memory
class
Worker
:
...
...
Prev
1
2
3
4
5
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment