Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
SIYIXNI
vllm
Commits
60dc62dc
"examples/community/latent_consistency_img2img.py" did not exist on "fcb2ec8c2f88c7519494bc679e59d8fc681cb65d"
Unverified
Commit
60dc62dc
authored
Dec 04, 2023
by
Roy
Committed by
GitHub
Dec 03, 2023
Browse files
add custom server params (#1868)
parent
0f90effc
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
9 additions
and
0 deletions
+9
-0
vllm/entrypoints/openai/api_server.py
vllm/entrypoints/openai/api_server.py
+4
-0
vllm/entrypoints/openai/protocol.py
vllm/entrypoints/openai/protocol.py
+4
-0
vllm/sampling_params.py
vllm/sampling_params.py
+1
-0
No files found.
vllm/entrypoints/openai/api_server.py
View file @
60dc62dc
...
...
@@ -253,8 +253,10 @@ async def create_chat_completion(request: ChatCompletionRequest,
n
=
request
.
n
,
presence_penalty
=
request
.
presence_penalty
,
frequency_penalty
=
request
.
frequency_penalty
,
repetition_penalty
=
request
.
repetition_penalty
,
temperature
=
request
.
temperature
,
top_p
=
request
.
top_p
,
min_p
=
request
.
min_p
,
stop
=
request
.
stop
,
stop_token_ids
=
request
.
stop_token_ids
,
max_tokens
=
request
.
max_tokens
,
...
...
@@ -497,9 +499,11 @@ async def create_completion(request: CompletionRequest, raw_request: Request):
best_of
=
request
.
best_of
,
presence_penalty
=
request
.
presence_penalty
,
frequency_penalty
=
request
.
frequency_penalty
,
repetition_penalty
=
request
.
repetition_penalty
,
temperature
=
request
.
temperature
,
top_p
=
request
.
top_p
,
top_k
=
request
.
top_k
,
min_p
=
request
.
min_p
,
stop
=
request
.
stop
,
stop_token_ids
=
request
.
stop_token_ids
,
ignore_eos
=
request
.
ignore_eos
,
...
...
vllm/entrypoints/openai/protocol.py
View file @
60dc62dc
...
...
@@ -75,6 +75,8 @@ class ChatCompletionRequest(BaseModel):
spaces_between_special_tokens
:
Optional
[
bool
]
=
True
add_generation_prompt
:
Optional
[
bool
]
=
True
echo
:
Optional
[
bool
]
=
False
repetition_penalty
:
Optional
[
float
]
=
1.0
min_p
:
Optional
[
float
]
=
0.0
class
CompletionRequest
(
BaseModel
):
...
...
@@ -102,6 +104,8 @@ class CompletionRequest(BaseModel):
stop_token_ids
:
Optional
[
List
[
int
]]
=
Field
(
default_factory
=
list
)
skip_special_tokens
:
Optional
[
bool
]
=
True
spaces_between_special_tokens
:
Optional
[
bool
]
=
True
repetition_penalty
:
Optional
[
float
]
=
1.0
min_p
:
Optional
[
float
]
=
0.0
class
LogProbs
(
BaseModel
):
...
...
vllm/sampling_params.py
View file @
60dc62dc
...
...
@@ -149,6 +149,7 @@ class SamplingParams:
# Zero temperature means greedy sampling.
self
.
top_p
=
1.0
self
.
top_k
=
-
1
self
.
min_p
=
0.0
self
.
_verify_greedy_sampling
()
def
_verify_args
(
self
)
->
None
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment