Skip to content

Commit 11604f0

Browse files
committed
Try 4 GPU test
1 parent 1294396 commit 11604f0

File tree

2 files changed

+5
-5
lines changed

2 files changed

+5
-5
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -113,6 +113,7 @@ steps:
113113
commands:
114114
- pytest -v -s distributed/test_pynccl.py
115115
- pytest -v -s spec_decode/e2e/test_integration_dist_tp4.py
116+
- pytest models/encoder_decoder/vision_language/test_broadcast.py -v -s -m distributed_4_gpus
116117

117118
- label: Metrics, Tracing Test # 10min
118119
num_gpus: 2
@@ -333,13 +334,13 @@ steps:
333334
- label: Custom Models Test
334335
#mirror_hardwares: [amd]
335336
working_dir: "/vllm-workspace/tests"
336-
num_gpus: 2
337+
num_gpus: 4
337338
optional: true
338339
commands:
339340
# PR authors can temporarily add commands below to test individual models
340341
# e.g. pytest -v -s models/encoder_decoder/vision_language/test_mllama.py
341342
# *To avoid merge conflicts, remember to REMOVE (not just comment out) them before merging the PR*
342-
- pytest -v -s models/encoder_decoder/vision_language/test_broadcast.py -v -s -m distributed_2_gpus
343+
- pytest -v -s models/encoder_decoder/vision_language/test_broadcast.py -v -s -m distributed_4_gpus
343344

344345
##### 1 GPU test #####
345346
##### multi gpus test #####
@@ -391,7 +392,6 @@ steps:
391392
# Avoid importing model tests that cause CUDA reinitialization error
392393
- pytest models/encoder_decoder/language/test_bart.py -v -s -m distributed_2_gpus
393394
- pytest models/decoder_only/vision_language/test_broadcast.py -v -s -m distributed_2_gpus
394-
- pytest models/encoder_decoder/vision_language/test_broadcast.py -v -s -m distributed_2_gpus
395395
- pytest -v -s spec_decode/e2e/test_integration_dist_tp2.py
396396
- pip install -e ./plugins/vllm_add_dummy_model
397397
- pytest -v -s distributed/test_distributed_oot.py

tests/models/encoder_decoder/vision_language/test_broadcast.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
from ....utils import multi_gpu_test
44

55

6-
@multi_gpu_test(num_gpus=2)
6+
@multi_gpu_test(num_gpus=4)
77
@pytest.mark.parametrize("distributed_executor_backend", ["ray", "mp"])
88
@pytest.mark.parametrize("model", [
99
"meta-llama/Llama-3.2-11B-Vision-Instruct",
@@ -14,7 +14,7 @@ def test_models(hf_runner, vllm_runner, image_assets,
1414
dtype = "half"
1515
max_tokens = 5
1616
num_logprobs = 5
17-
tensor_parallel_size = 2
17+
tensor_parallel_size = 4
1818

1919
if model.startswith("meta-llama/Llama-3.2-11B-Vision-Instruct"):
2020
from .test_mllama import models, run_test

0 commit comments

Comments
 (0)