Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
norm
vllm
Commits
c664b0e6
"docs/vscode:/vscode.git/clone" did not exist on "e17c26343df156c78de7887d95566c36624d2e35"
Unverified
Commit
c664b0e6
authored
Feb 01, 2024
by
zspo
Committed by
GitHub
Jan 31, 2024
Browse files
fix some bugs (#2689)
parent
d69ff0cb
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
8 additions
and
3 deletions
+8
-3
vllm/config.py
vllm/config.py
+4
-1
vllm/engine/async_llm_engine.py
vllm/engine/async_llm_engine.py
+4
-2
No files found.
vllm/config.py
View file @
c664b0e6
...
...
@@ -355,6 +355,9 @@ class ParallelConfig:
worker_use_ray: Whether to use Ray for model workers. Will be set to
True if either pipeline_parallel_size or tensor_parallel_size is
greater than 1.
max_parallel_loading_workers: Maximum number of multiple batches
when load model sequentially. To avoid RAM OOM when using tensor
parallel and large models.
disable_custom_all_reduce: Disable the custom all-reduce kernel and
fall back to NCCL.
"""
...
...
@@ -470,7 +473,7 @@ class LoRAConfig:
elif
self
.
max_cpu_loras
<
self
.
max_loras
:
raise
ValueError
(
f
"max_cpu_loras (
{
self
.
max_cpu_loras
}
) must be >= "
f
"max_
num_seq
s (
{
self
.
max_loras
}
)"
)
f
"max_
lora
s (
{
self
.
max_loras
}
)"
)
def
verify_with_model_config
(
self
,
model_config
:
ModelConfig
):
if
self
.
lora_dtype
in
(
None
,
"auto"
):
...
...
vllm/engine/async_llm_engine.py
View file @
c664b0e6
...
...
@@ -296,6 +296,8 @@ class AsyncLLMEngine:
async frontend will be executed in a separate process as the
model workers.
log_requests: Whether to log the requests.
max_log_len: Maximum number of prompt characters or prompt ID numbers
being printed in log.
start_engine_loop: If True, the background task to run the engine
will be automatically started in the generate call.
*args: Arguments for LLMEngine.
...
...
@@ -431,8 +433,8 @@ class AsyncLLMEngine:
logger
.
info
(
f
"Received request
{
request_id
}
: "
f
"prompt:
{
shortened_prompt
!
r
}
, "
f
"prefix_pos:
{
prefix_pos
}
,"
f
"sampling
params:
{
sampling_params
}
, "
f
"prompt
token
ids:
{
shortened_token_ids
}
, "
f
"sampling
_
params:
{
sampling_params
}
, "
f
"prompt
_
token
_
ids:
{
shortened_token_ids
}
, "
f
"lora_request:
{
lora_request
}
."
)
if
not
self
.
is_running
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment