2525from vllm .logger import init_logger
2626from vllm .tokenizers import TokenizerLike
2727from vllm .transformers_utils .processor import cached_processor_from_config
28- from vllm .transformers_utils .tokenizer import decode_tokens , encode_tokens
2928from vllm .utils .collection_utils import flatten_2d_lists , full_groupby
3029from vllm .utils .func_utils import get_allowed_kwarg_only_overrides
3130from vllm .utils .jsontree import JSONTree , json_map_leaves
@@ -80,21 +79,19 @@ def _cached_encode(
8079 tokenizer : TokenizerLike ,
8180 text : str ,
8281 * ,
83- add_special_tokens : bool | None = None ,
82+ add_special_tokens : bool = True ,
8483) -> list [int ]:
85- return encode_tokens ( tokenizer , text , add_special_tokens = add_special_tokens )
84+ return tokenizer . encode ( text , add_special_tokens = add_special_tokens )
8685
8786
8887@lru_cache (maxsize = 2048 )
8988def _cached_decode (
9089 tokenizer : TokenizerLike ,
9190 token_ids : tuple [int , ...],
9291 * ,
93- skip_special_tokens : bool | None = None ,
92+ skip_special_tokens : bool = False ,
9493) -> str :
95- return decode_tokens (
96- tokenizer , list (token_ids ), skip_special_tokens = skip_special_tokens
97- )
94+ return tokenizer .decode (list (token_ids ), skip_special_tokens = skip_special_tokens )
9895
9996
10097def _seq2text (
@@ -110,7 +107,7 @@ def _seq2text(
110107 raise ValueError ("You cannot decode tokens when `skip_tokenizer_init=True`" )
111108
112109 if not use_cache :
113- return decode_tokens ( tokenizer , seq )
110+ return tokenizer . decode ( seq )
114111
115112 return _cached_decode (tokenizer , tuple (seq ))
116113
@@ -126,7 +123,7 @@ def _seq2tokens(
126123 raise ValueError ("You cannot encode text when `skip_tokenizer_init=True`" )
127124
128125 if not use_cache :
129- return encode_tokens ( tokenizer , seq , add_special_tokens = False )
126+ return tokenizer . encode ( seq , add_special_tokens = False )
130127
131128 return _cached_encode (tokenizer , seq , add_special_tokens = False )
132129
@@ -2198,8 +2195,8 @@ def _get_enc_dec_inputs(
21982195 tokenizer = self .info .get_tokenizer ()
21992196 decoder_prompt_raw = self .create_decoder_prompt (prompt , mm_data )
22002197 if isinstance (decoder_prompt_raw , str ):
2201- decoder_prompt_ids = encode_tokens (
2202- tokenizer , decoder_prompt_raw , add_special_tokens = False
2198+ decoder_prompt_ids = tokenizer . encode (
2199+ decoder_prompt_raw , add_special_tokens = False
22032200 )
22042201 else :
22052202 decoder_prompt_ids = decoder_prompt_raw
0 commit comments