diff --git a/tests/unit/llms/test_llm_openai.py b/tests/unit/llms/test_llm_openai.py
index 6c01b2bd292ca4f3e4536cb57813858b92314227..c8fdde4a43b7fffa83500ec586e7b127d1190fe8 100644
--- a/tests/unit/llms/test_llm_openai.py
+++ b/tests/unit/llms/test_llm_openai.py
@@ -2,7 +2,10 @@ import pytest
 
 from semantic_router.llms import OpenAILLM
 from semantic_router.schema import Message
-from semantic_router.utils.function_call import get_schema_openai, convert_param_type_to_json_type
+from semantic_router.utils.function_call import (
+    get_schema_openai,
+    convert_param_type_to_json_type,
+)
 
 
 @pytest.fixture
@@ -168,7 +171,6 @@ class TestOpenAILLM:
             expected_error_message in actual_error_message
         ), f"Expected error message: '{expected_error_message}', but got: '{actual_error_message}'"
 
-    
     def test_convert_param_type_to_json_type(self):
         # Test conversion of basic types
         assert convert_param_type_to_json_type("int") == "number"
@@ -185,15 +187,22 @@ class TestOpenAILLM:
         function_schema = {"function": "get_user_data", "args": ["user_id"]}
 
         # Mock the __call__ method to return a JSON string as expected
-        mocker.patch.object(OpenAILLM, '__call__', return_value='{"user_id": "123"}')
+        mocker.patch.object(OpenAILLM, "__call__", return_value='{"user_id": "123"}')
         result = openai_llm.extract_function_inputs(query, function_schema)
 
         # Ensure the __call__ method is called with the correct parameters
         expected_messages = [
-            Message(role="system", content="You are an intelligent AI. Given a command or request from the user, call the function to complete the request."),
-            Message(role="user", content=query)
+            Message(
+                role="system",
+                content="You are an intelligent AI. Given a command or request from the user, call the function to complete the request.",
+            ),
+            Message(role="user", content=query),
         ]
-        openai_llm.__call__.assert_called_once_with(messages=expected_messages, function_schema=function_schema)
+        openai_llm.__call__.assert_called_once_with(
+            messages=expected_messages, function_schema=function_schema
+        )
 
         # Check if the result is as expected
-        assert result == {"user_id": "123"}, "The function inputs should match the expected dictionary."
\ No newline at end of file
+        assert result == {
+            "user_id": "123"
+        }, "The function inputs should match the expected dictionary."
diff --git a/tests/unit/test_route.py b/tests/unit/test_route.py
index 5b91f9d32770bf0b736eb34540dac4e8cc3c9248..3cedeadb998fa6632e87672752b9f6da9ddfe4c6 100644
--- a/tests/unit/test_route.py
+++ b/tests/unit/test_route.py
@@ -5,6 +5,7 @@ import pytest
 from semantic_router.llms import BaseLLM
 from semantic_router.route import Route, is_valid
 
+
 # Is valid test:
 def test_is_valid_with_valid_json():
     valid_json = '{"name": "test_route", "utterances": ["hello", "hi"]}'