diff --git a/src/llama_recipes/finetuning.py b/src/llama_recipes/finetuning.py
index 4cec952ac4f1e2db55601f0a008a770eecfd70a6..0e140a7971115b77000ced3cd7d04cdecd5c0e64 100644
--- a/src/llama_recipes/finetuning.py
+++ b/src/llama_recipes/finetuning.py
@@ -167,7 +167,7 @@ def main(**kwargs):
         # Load the pre-trained peft model checkpoint and setup its configuration
         if train_config.from_peft_checkpoint:
             model = PeftModel.from_pretrained(model, train_config.from_peft_checkpoint, is_trainable=True)
-            peft_config = model.peft_config()
+            peft_config = model.peft_config
         # Generate the peft config and start fine-tuning from original model
         else:
             peft_config = generate_peft_config(train_config, kwargs)