2020
2121# Third Party
2222from transformers import (
23- AutoModelForCausalLM ,
24- AutoModelForVision2Seq ,
25- AutoProcessor ,
26- AutoTokenizer ,
23+ AutoModelForImageTextToText , # AutoModelForVision2Seq was renamed to this in transformers v5
2724)
25+ from transformers import AutoModelForCausalLM , AutoProcessor , AutoTokenizer
2826import torch
2927
3028# First Party
@@ -128,16 +126,17 @@ def test_special_tokens_before_and_after():
128126 model = AutoModelForCausalLM .from_pretrained (MODEL_NAME )
129127
130128 input_tokenizer_len = len (tokenizer .get_vocab ())
131- addn_spl_tokens_before = tokenizer .special_tokens_map .get (
132- "additional_special_tokens"
133- )
129+ addn_spl_tokens_before = list (tokenizer .extra_special_tokens )
134130 assert (
135131 len (addn_spl_tokens_before ) > 0
136132 ), "this test needs tokenizer special tokens to not be empty before testing"
137133
138134 special_tokens_dict = {"sep_token" : "<SEP>" , "pad_token" : "<PAD>" }
139135 addn_spl_tokens_added = ["<NotSeenTokenA>" , "<NotSeenTokenB>" , "<NotSeenTokenC>" ]
140- special_tokens_dict ["additional_special_tokens" ] = addn_spl_tokens_added
136+ # for transformers v5: merge existing extra_special_tokens with new ones to prevent replacement
137+ special_tokens_dict ["additional_special_tokens" ] = (
138+ list (tokenizer .extra_special_tokens ) + addn_spl_tokens_added
139+ )
141140
142141 resize_result = tokenizer_and_embedding_resize (
143142 special_tokens_dict = special_tokens_dict ,
@@ -150,9 +149,7 @@ def test_special_tokens_before_and_after():
150149 addn_spl_tokens_before .extend (addn_spl_tokens_added )
151150 expected_addn_special_tokens = addn_spl_tokens_before
152151 expected_embedding_size = input_tokenizer_len + len (addn_spl_tokens_added ) + 2
153- addn_spl_tokens_after = tokenizer .special_tokens_map .get (
154- "additional_special_tokens"
155- )
152+ addn_spl_tokens_after = list (tokenizer .extra_special_tokens )
156153
157154 assert "<SEP>" in tokenizer .get_vocab ()
158155 assert "<PAD>" in tokenizer .get_vocab ()
@@ -212,7 +209,9 @@ def test_resize_with_multiple_of():
212209
213210
214211def test_resize_llama_vision_model ():
215- model = AutoModelForVision2Seq .from_pretrained (TINY_LLAMA_VISION_MODEL_NAME )
212+ model = AutoModelForImageTextToText .from_pretrained (
213+ TINY_LLAMA_VISION_MODEL_NAME
214+ ) # AutoModelForVision2Seq was renamed to AutoModelForImageTextToText in transformers v5
216215 processor = AutoProcessor .from_pretrained (TINY_LLAMA_VISION_MODEL_NAME )
217216 tokenizer = processor .tokenizer
218217
0 commit comments