Skip to content

Commit cd8f907

Browse files
committed
fix compatibility
1 parent 7d75ebb commit cd8f907

File tree

3 files changed

+3
-3
lines changed

3 files changed

+3
-3
lines changed

vllm/compilation/mirage_backend.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -331,7 +331,6 @@ def compile_or_call(*args):
331331

332332
self.compiled = True
333333

334-
logger.info(f"[Mirage] Calling the compiled result...")
335334
result_hidden_states = self.mpk()
336335

337336
return (result_hidden_states,)

vllm/platforms/cuda.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -346,7 +346,7 @@ def get_attn_backend_cls(
346346
FLASH_ATTN_V1 = "vllm.v1.attention.backends.flash_attn.FlashAttentionBackend" # noqa: E501
347347
TREE_ATTN_V1 = "vllm.v1.attention.backends.tree_attn.TreeAttentionBackend" # noqa: E501
348348
XFORMERS_V1 = "vllm.v1.attention.backends.xformers.XFormersAttentionBackend" # noqa: E501
349-
MIRAGE_V1 = "vllm.v1.attention.backends.mirage.MirageBackend" # noqa: E501
349+
MIRAGE_V1 = "vllm.v1.attention.backends.mirage.MirageAttentionBackend" # noqa: E501
350350

351351
use_fp8_kv_cache = kv_cache_dtype is not None and kv_cache_dtype.startswith(
352352
"fp8"

vllm/v1/attention/backends/mirage.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,8 @@
2222

2323
from vllm.platforms import current_platform
2424
from vllm.triton_utils import tl, triton
25-
from vllm.utils import cdiv, is_pin_memory_available
25+
from vllm.utils.math_utils import cdiv
26+
from vllm.utils.platform_utils import is_pin_memory_available
2627
from vllm.v1.attention.backends.utils import (
2728
AttentionCGSupport,
2829
AttentionMetadataBuilder,

0 commit comments

Comments
 (0)