File tree Expand file tree Collapse file tree 2 files changed +9
-4
lines changed Expand file tree Collapse file tree 2 files changed +9
-4
lines changed Original file line number Diff line number Diff line change 14
14
agents :
15
15
queue : cpu_queue_postmerge
16
16
commands :
17
- - " DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.6.3 --tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
17
+ - " DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.6.3 --build-arg torch_cuda_arch_list='7.0 7.5 8.0 8.9 9.0+PTX' -- tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
18
18
- " mkdir artifacts"
19
19
- " docker run --rm -v $(pwd)/artifacts:/artifacts_host vllm-ci:build-image bash -c 'cp -r dist /artifacts_host && chmod -R a+rw /artifacts_host'"
20
20
- " bash .buildkite/scripts/upload-wheels.sh"
31
31
agents :
32
32
queue : cpu_queue_postmerge
33
33
commands :
34
- - " DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=11.8.0 --tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
34
+ - " DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=11.8.0 --build-arg torch_cuda_arch_list='7.0 7.5 8.0 8.9 9.0+PTX' -- tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
35
35
- " mkdir artifacts"
36
36
- " docker run --rm -v $(pwd)/artifacts:/artifacts_host vllm-ci:build-image bash -c 'cp -r dist /artifacts_host && chmod -R a+rw /artifacts_host'"
37
37
- " bash .buildkite/scripts/upload-wheels.sh"
Original file line number Diff line number Diff line change @@ -257,8 +257,13 @@ RUN --mount=type=cache,target=/root/.cache/uv \
257
257
if [ "$TARGETPLATFORM" != "linux/arm64" ]; then \
258
258
# uv pip install --system https://github.com/flashinfer-ai/flashinfer/releases/download/v0.2.4/flashinfer_python-0.2.4+cu124torch2.6-cp38-abi3-linux_x86_64.whl ; \
259
259
# TESTING: install FlashInfer from source to test 2.7.0 final RC
260
- FLASHINFER_ENABLE_AOT=1 TORCH_CUDA_ARCH_LIST='7.5 8.0 8.9 9.0 10.0+PTX' \
261
- uv pip install --system --no-build-isolation "git+https://github.com/flashinfer-ai/flashinfer@e00e8cedbfcb220f328fd36aa8f529f869b01e6b" ; \
260
+ if [[ "$CUDA_VERSION" == 12.8* ]]; then \
261
+ export TORCH_CUDA_ARCH_LIST='7.5 8.0 8.9 9.0 10.0+PTX' ; \
262
+ else \
263
+ export TORCH_CUDA_ARCH_LIST='7.5 8.0 8.9 9.0+PTX' ; \
264
+ fi && \
265
+ export FLASHINFER_ENABLE_AOT=1; \
266
+ uv pip install --system --no-build-isolation "git+https://github.com/flashinfer-ai/flashinfer@21ea1d2545f74782b91eb8c08fd503ac4c0743fc" ; \
262
267
fi
263
268
COPY examples examples
264
269
COPY benchmarks benchmarks
You can’t perform that action at this time.
0 commit comments