Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
19 commits
Select commit Hold shift + click to select a range
ca00740
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
7dd1d55
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
713ae5b
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
82596db
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
e907f8d
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
b32a084
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
59c7e74
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
6ce992e
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
e4ef0d7
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
5f0caa1
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
a18a43b
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
9abf689
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
c068154
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
8d1bc67
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
3c83c7b
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
f6bdf83
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
8476677
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
470ad4c
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
171ce8c
[Frontend] supports deepseekv32 chat template
chaunceyjiang Dec 2, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion vllm/config/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@
"transcription",
"draft",
]
TokenizerMode = Literal["auto", "hf", "slow", "mistral"]
TokenizerMode = Literal["auto", "hf", "slow", "mistral", "deepseek_v32"]
ModelDType = Literal["auto", "half", "float16", "bfloat16", "float", "float32"]
LogprobsMode = Literal[
"raw_logits", "raw_logprobs", "processed_logits", "processed_logprobs"
Expand Down Expand Up @@ -141,6 +141,7 @@ class ModelConfig:
- "hf" will use the fast tokenizer if available.\n
- "slow" will always use the slow tokenizer.\n
- "mistral" will always use the tokenizer from `mistral_common`.\n
- "deepseek_v32" will always use the tokenizer from `deepseek_v32`.\n
- Other custom values can be supported via plugins."""
trust_remote_code: bool = False
"""Trust remote code (e.g., from HuggingFace) when downloading the model
Expand Down
9 changes: 8 additions & 1 deletion vllm/entrypoints/openai/serving_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@
from vllm.pooling_params import PoolingParams
from vllm.reasoning import ReasoningParser, ReasoningParserManager
from vllm.sampling_params import BeamSearchParams, SamplingParams
from vllm.tokenizers import MistralTokenizer, TokenizerLike
from vllm.tokenizers import DeepseekV32Tokenizer, MistralTokenizer, TokenizerLike
from vllm.tracing import (
contains_trace_headers,
extract_trace_headers,
Expand Down Expand Up @@ -1129,6 +1129,13 @@ async def _preprocess_chat(
messages=messages,
**_chat_template_kwargs,
)
elif isinstance(tokenizer, DeepseekV32Tokenizer):
request_prompt = tokenizer.apply_chat_template(
conversation=conversation,
messages=messages,
model_config=model_config,
**_chat_template_kwargs,
)
else:
request_prompt = apply_hf_chat_template(
tokenizer=tokenizer,
Expand Down
2 changes: 2 additions & 0 deletions vllm/tokenizers/__init__.py
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
# SPDX-License-Identifier: Apache-2.0
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project

from .deepseekv32 import DeepseekV32Tokenizer
from .hf import HfTokenizer
from .mistral import MistralTokenizer
from .protocol import TokenizerLike
Expand All @@ -21,4 +22,5 @@
"get_tokenizer",
"cached_tokenizer_from_config",
"init_tokenizer_from_config",
"DeepseekV32Tokenizer",
]
Loading