Skip to content
Snippets Groups Projects
Commit fcc817e9 authored by hongbo.mo's avatar hongbo.mo
Browse files

bugfix: remove duplicate load_peft_model

parent 03faba66
No related branches found
No related tags found
No related merge requests found
...@@ -99,11 +99,7 @@ def main( ...@@ -99,11 +99,7 @@ def main(
print("Skipping the inference as the prompt is not safe.") print("Skipping the inference as the prompt is not safe.")
sys.exit(1) # Exit the program with an error status sys.exit(1) # Exit the program with an error status
if peft_model: batch = tokenizer(user_prompt, padding='max_length', truncation=True, max_length=max_padding_length, return_tensors="pt")
model = load_peft_model(model, peft_model)
model.eval()
batch = tokenizer(user_prompt, padding='max_length', truncation=True,max_length=max_padding_length,return_tensors="pt")
batch = {k: v.to("cuda") for k, v in batch.items()} batch = {k: v.to("cuda") for k, v in batch.items()}
start = time.perf_counter() start = time.perf_counter()
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment