diff --git a/src/tests/test_finetuning.py b/src/tests/test_finetuning.py index b695577523bc75884950dd515cbffde316aed2ed..9d9e6faaced639b9a2589fec51d00bf8c85ff26e 100644 --- a/src/tests/test_finetuning.py +++ b/src/tests/test_finetuning.py @@ -246,8 +246,6 @@ def test_finetuning_weight_decay( args, kwargs = train.call_args optimizer = args[4] - print(optimizer.state_dict()) - assert isinstance(optimizer, AdamW) assert optimizer.state_dict()["param_groups"][0]["weight_decay"] == approx(0.01) diff --git a/src/tests/utils.py b/src/tests/utils.py index 9dbf30746a66061cc9d781c1ee38e4136bf75c74..14b96a9a1fe46251462ca74768baa65f6f9d402e 100644 --- a/src/tests/utils.py +++ b/src/tests/utils.py @@ -20,7 +20,6 @@ class FakeTokenizer(object): self.padding_side = "left" def __call__(self, *args, **kwargs): - print(f"{kwargs=}") ids = self.encode(*args, **kwargs) return {"input_ids": ids}