Skip to content

Commit 1e26077

Browse files
LiuXiaoxuanPKUgarg-amit
authored andcommitted
[CI][SpecDecode] Fix spec decode tests, use flash attention backend for spec decode CI tests. (vllm-project#8975)
Signed-off-by: Amit Garg <[email protected]>
1 parent bad2e20 commit 1e26077

File tree

2 files changed

+4
-3
lines changed

2 files changed

+4
-3
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -207,8 +207,6 @@ steps:
207207
- vllm/spec_decode
208208
- tests/spec_decode
209209
commands:
210-
# See https://github.com/vllm-project/vllm/issues/5152
211-
- export VLLM_ATTENTION_BACKEND=XFORMERS
212210
- pytest -v -s spec_decode/e2e/test_multistep_correctness.py
213211
- pytest -v -s spec_decode --ignore=spec_decode/e2e/test_multistep_correctness.py
214212

tests/spec_decode/test_multi_step_worker.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -673,7 +673,10 @@ def test_use_draft_model_runner_advance_step():
673673
worker.model_runner._gpu_advance_step.side_effect = ValueError(
674674
exception_secret)
675675

676-
seq_group_metadata_list, _, _ = create_batch(batch_size, k)
676+
seq_group_metadata_list, _, _ = create_batch(batch_size,
677+
k,
678+
block_size=block_size,
679+
num_gpu_blocks=num_gpu_blocks)
677680

678681
# Fallback (should not call) when num_steps=1.
679682
execute_model_req = ExecuteModelRequest(

0 commit comments

Comments
 (0)