diff --git a/semantic_router/llms/openai.py b/semantic_router/llms/openai.py
index 4e2af247a85e213675580ca71db6161a6f034455..9072821092c32bb105865624b051890d21c48757 100644
--- a/semantic_router/llms/openai.py
+++ b/semantic_router/llms/openai.py
@@ -44,12 +44,6 @@ class OpenAILLM(BaseLLM):
                 tools = [function_schema] 
             else:
                 tools = None
-            # DEBUGGING: Start.
-            print('#'*50)
-            print('tools')
-            print(tools)
-            print('#'*50)
-            # DEBUGGING: End.
             completion = self.client.chat.completions.create(
                 model=self.name,
                 messages=[m.to_openai() for m in messages],
@@ -59,14 +53,6 @@ class OpenAILLM(BaseLLM):
             )
 
             output = completion.choices[0].message.content
-            # DEBUGGING: Start.
-            print('#'*50)
-            # print('print(completion.choices[0].message.function_call)')
-            # print(print(completion.choices[0].message.function_call))
-            print('completion.choices[0].message.tool_calls')
-            print(completion.choices[0].message.tool_calls)
-            print('#'*50)
-            # DEBUGGING: End.
 
             if function_schema:
                 return completion.choices[0].message.tool_calls
@@ -93,24 +79,10 @@ class OpenAILLM(BaseLLM):
         output = self(messages=messages, function_schema=function_schema)
         if not output:
             raise Exception("No output generated for extract function input")
-        # DEBUGGING: Start.
-        print('#'*50)
-        print('output')
-        print(output)
-        print('#'*50)
-        # DEBUGGING: End.
         if len(output) != 1:
             raise ValueError("Invalid output, expected a single tool to be called")
         tool_call = output[0]
         arguments_json = tool_call.function.arguments
         function_inputs = json.loads(arguments_json)
-
-        # DEBUGGING: Start.
-        print('#'*50)
-        print('function_inputs')
-        print(function_inputs)
-        print('#'*50)
-        # DEBUGGING: End.
-
         return function_inputs
                 
\ No newline at end of file
diff --git a/semantic_router/route.py b/semantic_router/route.py
index b0580885de33365188bdcc111b55bdd73ae2e3ee..3a62fab163ca33737cdcc170e42d2dfa5120295e 100644
--- a/semantic_router/route.py
+++ b/semantic_router/route.py
@@ -76,12 +76,6 @@ class Route(BaseModel):
             extracted_inputs = self.llm.extract_function_inputs(
                 query=query, function_schema=self.function_schema
             )
-            # DEBUGGING: Start.
-            print('#'*50)
-            print('extracted_inputs')
-            print(extracted_inputs)
-            print('#'*50)
-            # DEBUGGING: End.
             func_call = extracted_inputs
         elif self.openai_function_schema:
             if not isinstance(self.llm, OpenAILLM):