diff --git a/src/tests/test_finetuning.py b/src/tests/test_finetuning.py
index b695577523bc75884950dd515cbffde316aed2ed..9d9e6faaced639b9a2589fec51d00bf8c85ff26e 100644
--- a/src/tests/test_finetuning.py
+++ b/src/tests/test_finetuning.py
@@ -246,8 +246,6 @@ def test_finetuning_weight_decay(
     args, kwargs = train.call_args
     optimizer = args[4]
 
-    print(optimizer.state_dict())
-
     assert isinstance(optimizer, AdamW)
     assert optimizer.state_dict()["param_groups"][0]["weight_decay"] == approx(0.01)
 
diff --git a/src/tests/utils.py b/src/tests/utils.py
index 9dbf30746a66061cc9d781c1ee38e4136bf75c74..14b96a9a1fe46251462ca74768baa65f6f9d402e 100644
--- a/src/tests/utils.py
+++ b/src/tests/utils.py
@@ -20,7 +20,6 @@ class FakeTokenizer(object):
         self.padding_side = "left"
 
     def __call__(self, *args, **kwargs):
-        print(f"{kwargs=}")
         ids = self.encode(*args, **kwargs)
         return {"input_ids": ids}