Unverified Commit 20044cab authored by Lily Liu's avatar Lily Liu Committed by GitHub
Browse files

Fix log message in scheduler (#652)

parent 64f23c29
...@@ -190,13 +190,13 @@ class Scheduler: ...@@ -190,13 +190,13 @@ class Scheduler:
break break
num_prompt_tokens = seq_group.get_seqs()[0].get_len() num_prompt_tokens = seq_group.get_seqs()[0].get_len()
if num_prompt_tokens > min( prompt_limit = min(
self.scheduler_config.max_model_len, self.scheduler_config.max_model_len,
self.scheduler_config.max_num_batched_tokens): self.scheduler_config.max_num_batched_tokens)
if num_prompt_tokens > prompt_limit:
logger.warning( logger.warning(
f"Input prompt ({num_prompt_tokens} tokens) is too long" f"Input prompt ({num_prompt_tokens} tokens) is too long"
" and exceeds limit of " f" and exceeds limit of {prompt_limit}")
f"{self.scheduler_config.max_model_len}")
for seq in seq_group.get_seqs(): for seq in seq_group.get_seqs():
seq.status = SequenceStatus.FINISHED_IGNORED seq.status = SequenceStatus.FINISHED_IGNORED
ignored_seq_groups.append(seq_group) ignored_seq_groups.append(seq_group)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment