From 26003a478f8e7aed568effb062d2f88b070a2dc1 Mon Sep 17 00:00:00 2001 From: Sanjay Nadhavajhala Date: Fri, 28 Jun 2024 16:19:13 -0700 Subject: [PATCH] add an error handling and remove prints --- vllm/entrypoints/openai/serving_chat.py | 19 ++++++++++--------- 1 file changed, 10 insertions(+), 9 deletions(-) diff --git a/vllm/entrypoints/openai/serving_chat.py b/vllm/entrypoints/openai/serving_chat.py index 77c26c898e7f9..9d07495dc7d54 100644 --- a/vllm/entrypoints/openai/serving_chat.py +++ b/vllm/entrypoints/openai/serving_chat.py @@ -509,7 +509,6 @@ async def chat_completion_full_generator( choices: List[ChatCompletionResponseChoice] = [] role = self.get_chat_request_role(request) - print("========================output========================") for output in final_res.outputs: token_ids = output.token_ids out_logprobs = output.logprobs @@ -524,8 +523,6 @@ async def chat_completion_full_generator( else: logprobs = None - # TODO: use llama_tools to parse the output.text - print(output) finish_reason = output.finish_reason if request.tool_choice and type( @@ -546,12 +543,16 @@ async def chat_completion_full_generator( tool_calls = [] if function_output: print(f"Parsed function output: {function_output}\n\n") - for fc in function_output: - function = FunctionCall(name=fc["function"]["name"], arguments=fc["function"]["arguments"]) - call = ToolCall(function=function) - tool_calls.append(call) - content = "" - finish_reason = "tool_calls" + try: + for fc in function_output: + function = FunctionCall(name=fc["function"]["name"], arguments=fc["function"]["arguments"]) + call = ToolCall(function=function) + tool_calls.append(call) + content = "" + finish_reason = "tool_calls" + except Exception as e: + content = str(function_output) + print(f"Error extract functions from output: {e}") message = ChatMessage(role=role, content=content, tool_calls=tool_calls) choice_data = ChatCompletionResponseChoice(