From 8ac44ef3bee1ed09f3dd4ad45b6f2769a4b9f0c8 Mon Sep 17 00:00:00 2001 From: Matthias Reso <13337103+mreso@users.noreply.github.com> Date: Thu, 14 Sep 2023 14:26:23 +0000 Subject: [PATCH] Fix vocab size mismatch in inference due to added pad token --- examples/inference.py | 8 +------- 1 file changed, 1 insertion(+), 7 deletions(-) diff --git a/examples/inference.py b/examples/inference.py index 28952920..ab4e7139 100644 --- a/examples/inference.py +++ b/examples/inference.py @@ -72,13 +72,7 @@ def main( print("Module 'optimum' not found. Please install 'optimum' it before proceeding.") tokenizer = LlamaTokenizer.from_pretrained(model_name) - tokenizer.add_special_tokens( - { - - "pad_token": "<PAD>", - } - ) - model.resize_token_embeddings(model.config.vocab_size + 1) + tokenizer.pad_token = tokenizer.eos_token safety_checker = get_safety_checker(enable_azure_content_safety, enable_sensitive_topics, -- GitLab