From 61aff1fd8e4c154342c8515e6146b01239306fe6 Mon Sep 17 00:00:00 2001 From: liuzhenghua-jk Date: Thu, 6 Jun 2024 11:58:29 +0800 Subject: [PATCH] fix: extra newline token in stream mode --- basic_demo/openai_api_server.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/basic_demo/openai_api_server.py b/basic_demo/openai_api_server.py index f03b120..07ea206 100644 --- a/basic_demo/openai_api_server.py +++ b/basic_demo/openai_api_server.py @@ -484,7 +484,8 @@ async def predict_stream(model_id, gen_params): object="chat.completion.chunk" ) yield "{}".format(chunk.model_dump_json(exclude_unset=True)) - + if not has_send_first_chunk and output.startswith("\n"): + output = output[1:] send_msg = delta_text if has_send_first_chunk else output has_send_first_chunk = True message = DeltaMessage(