[CI/Build] Fix disabled v1 attention backend selection test (#25471)

Signed-off-by: Isotr0py <mozf@mail2.sysu.edu.cn>
This commit is contained in:
Isotr0py
2025-09-23 21:05:46 +08:00
committed by GitHub
parent 0d9fe260dd
commit b6a136b58c

View File

@ -67,7 +67,6 @@ def generate_params():
return params
@pytest.mark.skip(reason="Skipped for now. Should be revisited.")
@pytest.mark.parametrize("device, name, use_mla, block_size",
generate_params())
def test_env(
@ -189,7 +188,7 @@ def test_env(
# FlashMLA only supports block_size == 64
pytest.skip("FlashMLA only supports block_size 64")
else:
from vllm.attention.backends.flashmla import (
from vllm.v1.attention.backends.mla.flashmla import ( # noqa: E501
is_flashmla_supported)
is_supported, _ = is_flashmla_supported()
if not is_supported: