Skip to content

Commit 0f3ce4c

Browse files
yma11jikunshang
andauthored
[XPU] Fix spec-decode UTs under tests/v1/spec_decode (#38491)
Signed-off-by: Yan Ma <yan.ma@intel.com> Co-authored-by: Kunshang Ji <kunshang.ji@intel.com>
1 parent af661a1 commit 0f3ce4c

3 files changed

Lines changed: 2 additions & 14 deletions

File tree

tests/v1/spec_decode/test_eagle.py

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -755,12 +755,6 @@ def test_load_model(
755755
use_distinct_lm_head,
756756
monkeypatch,
757757
):
758-
if attn_backend == "TRITON_ATTN" and not current_platform.is_rocm():
759-
pytest.skip(
760-
"TRITON_ATTN does not support "
761-
"multi-token eagle spec decode on current platform"
762-
)
763-
764758
if attn_backend == "ROCM_AITER_FA" and current_platform.is_rocm():
765759
monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1")
766760

tests/v1/spec_decode/test_eagle_step_kernel.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,8 +15,8 @@
1515

1616
# Skip if no CUDA - Triton kernel requires GPU
1717
pytest.importorskip("triton")
18-
if not torch.cuda.is_available():
19-
pytest.skip("CUDA required for EAGLE kernel tests", allow_module_level=True)
18+
if not current_platform.is_cuda_alike() and not current_platform.is_xpu():
19+
pytest.skip("CUDA/XPU required for EAGLE kernel tests", allow_module_level=True)
2020

2121

2222
def _reference_eagle_step_slot_mapping(

tests/v1/spec_decode/test_max_len.py

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -38,12 +38,6 @@ def test_ngram_max_len(num_speculative_tokens: int):
3838
def test_eagle_max_len(
3939
monkeypatch: pytest.MonkeyPatch, num_speculative_tokens: int, attn_backend: str
4040
):
41-
if attn_backend == "TRITON_ATTN" and not current_platform.is_rocm():
42-
pytest.skip(
43-
"TRITON_ATTN does not support "
44-
"multi-token eagle spec decode on current platform"
45-
)
46-
4741
if attn_backend == "ROCM_AITER_FA" and current_platform.is_rocm():
4842
monkeypatch.setenv("VLLM_ROCM_USE_AITER", "1")
4943

0 commit comments

Comments
 (0)