diff --git a/semantic_router/llms/openai.py b/semantic_router/llms/openai.py index 7df79da8b5163d6ecd2ee21b7032301a1f30b44d..2a5311958f91a071212b6b14199205cddea3f01a 100644 --- a/semantic_router/llms/openai.py +++ b/semantic_router/llms/openai.py @@ -101,8 +101,7 @@ class OpenAILLM(BaseLLM): content = completion.choices[0].message.content if content is None: raise ValueError("Invalid output, expected content.") - output = str(content) # str in keepign with base type. - + output = content return output except Exception as e: @@ -112,10 +111,11 @@ class OpenAILLM(BaseLLM): def extract_function_inputs( self, query: str, function_schemas: List[Dict[str, Any]] ) -> List[Dict[str, Any]]: - messages = [] system_prompt = "You are an intelligent AI. Given a command or request from the user, call the function to complete the request." - messages.append(Message(role="system", content=system_prompt)) - messages.append(Message(role="user", content=query)) + messages = [ + Message(role="system", content=system_prompt), + Message(role="user", content=query), + ] output = self(messages=messages, function_schemas=function_schemas) if not output: raise Exception("No output generated for extract function input")