"...git@developer.sourcefind.cn:chenpangpang/open-webui.git" did not exist on "797ed9ac7f5590bd0358e55f632ec5f729873d21"
Commit 2e0fa1c6 authored by Michael Poluektov's avatar Michael Poluektov
Browse files

refac: rename stream_message_template

parent f8726719
...@@ -20,7 +20,7 @@ from apps.webui.models.models import Models ...@@ -20,7 +20,7 @@ from apps.webui.models.models import Models
from apps.webui.utils import load_function_module_by_id from apps.webui.utils import load_function_module_by_id
from utils.misc import ( from utils.misc import (
stream_message_template, openai_chat_chunk_message_template,
openai_chat_completion_message_template, openai_chat_completion_message_template,
add_or_update_system_message, add_or_update_system_message,
) )
...@@ -227,7 +227,7 @@ def process_line(form_data: dict, line): ...@@ -227,7 +227,7 @@ def process_line(form_data: dict, line):
if line.startswith("data:"): if line.startswith("data:"):
return f"{line}\n\n" return f"{line}\n\n"
else: else:
line = stream_message_template(form_data["model"], line) line = openai_chat_chunk_message_template(form_data["model"], line)
return f"data: {json.dumps(line)}\n\n" return f"data: {json.dumps(line)}\n\n"
...@@ -371,7 +371,9 @@ async def generate_function_chat_completion(form_data, user): ...@@ -371,7 +371,9 @@ async def generate_function_chat_completion(form_data, user):
return return
if isinstance(res, str): if isinstance(res, str):
message = stream_message_template(form_data["model"], res) message = openai_chat_chunk_message_template(
form_data["model"], res
)
yield f"data: {json.dumps(message)}\n\n" yield f"data: {json.dumps(message)}\n\n"
if isinstance(res, Iterator): if isinstance(res, Iterator):
...@@ -383,7 +385,9 @@ async def generate_function_chat_completion(form_data, user): ...@@ -383,7 +385,9 @@ async def generate_function_chat_completion(form_data, user):
yield process_line(form_data, line) yield process_line(form_data, line)
if isinstance(res, str) or isinstance(res, Generator): if isinstance(res, str) or isinstance(res, Generator):
finish_message = stream_message_template(form_data["model"], "") finish_message = openai_chat_chunk_message_template(
form_data["model"], ""
)
finish_message["choices"][0]["finish_reason"] = "stop" finish_message["choices"][0]["finish_reason"] = "stop"
yield f"data: {json.dumps(finish_message)}\n\n" yield f"data: {json.dumps(finish_message)}\n\n"
yield "data: [DONE]" yield "data: [DONE]"
......
...@@ -97,7 +97,7 @@ def message_template(model: str): ...@@ -97,7 +97,7 @@ def message_template(model: str):
} }
def stream_message_template(model: str, message: str): def openai_chat_chunk_message_template(model: str, message: str):
template = message_template(model) template = message_template(model)
template["object"] = "chat.completion.chunk" template["object"] = "chat.completion.chunk"
template["choices"][0]["delta"] = {"content": message} template["choices"][0]["delta"] = {"content": message}
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment