Compare commits

...

27 Commits

Author SHA1 Message Date
2f10f1b888 add config 2025-10-02 12:49:06 -07:00
517f267085 add config 2025-10-02 12:47:06 -07:00
cac7242b91 add config 2025-10-02 10:42:53 -07:00
b54dc58cb5 add config 2025-10-02 09:38:22 -07:00
4efdd216bd add config 2025-10-02 09:36:24 -07:00
95654a32f5 add config 2025-10-02 09:35:34 -07:00
2f5c2ccf7a add config 2025-10-02 09:35:18 -07:00
813cae6074 add config 2025-10-02 09:34:38 -07:00
ef4730d5bb add config 2025-10-02 08:34:27 -07:00
3ad3df90c3 add config 2025-10-02 08:33:04 -07:00
257bf0e654 add config 2025-10-02 08:31:23 -07:00
02d16522d8 add config 2025-10-02 08:29:56 -07:00
e6d3372157 Would this work?
Why did it link with CUDA runtime here https://github.com/pytorch/pytorch/actions/runs/18187580810/job/51775137099?pr=164361#step:17:20314?
2025-10-02 03:21:49 -07:00
3f3d86adf2 Bring in https://github.com/vllm-project/vllm/pull/25730 2025-10-02 01:24:01 -07:00
58478b0ab8 Another tweak 2025-10-02 00:07:11 -07:00
98e554222f [no ci] Another tweak 2025-10-02 00:06:15 -07:00
700d608f4a Remove somewhat unnecessary change 2025-10-01 22:48:01 -07:00
1b27857415 Minor tweak 2025-10-01 22:44:50 -07:00
73995b1b5e add config 2025-10-01 18:04:39 -07:00
03d7c77071 add config 2025-10-01 16:07:45 -07:00
019d9cda40 add config 2025-10-01 15:53:13 -07:00
3620191a0a add config 2025-10-01 15:47:36 -07:00
5a722ca130 add config 2025-10-01 14:00:05 -07:00
8746e3cea2 add config 2025-10-01 12:45:52 -07:00
8cd1996b57 add config 2025-10-01 11:22:17 -07:00
73c23f3554 add config 2025-10-01 09:49:30 -07:00
1da3d6f595 add config 2025-10-01 08:32:45 -07:00
3 changed files with 19 additions and 4 deletions

View File

@ -202,7 +202,7 @@ ARG max_jobs=16
ENV MAX_JOBS=${max_jobs}
ARG nvcc_threads=4
ENV NVCC_THREADS=$nvcc_threads
ARG torch_cuda_arch_list='8.0;8.6;8.9;9.0'
ARG torch_cuda_arch_list='8.0 8.6 8.9 9.0'
ENV TORCH_CUDA_ARCH_LIST=${torch_cuda_arch_list}
ARG USE_SCCACHE
@ -297,16 +297,28 @@ RUN echo "[INFO] Listing current directory before torch install step:" && \
echo "[INFO] Showing torch_build_versions.txt content:" && \
cat torch_build_versions.txt
# Install build and runtime dependencies, this is needed for flashinfer install
COPY requirements/build.txt requirements/build.txt
COPY use_existing_torch.py use_existing_torch.py
RUN python3 use_existing_torch.py
RUN cat requirements/build.txt
# Install uv for faster pip installs if not existed
RUN --mount=type=cache,target=/root/.cache/uv \
if ! python3 -m uv --version > /dev/null 2>&1; then \
python3 -m pip install uv==0.8.4; \
fi
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"
# Use copy mode to avoid hardlink failures with Docker cache mounts
ENV UV_LINK_MODE=copy
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -r requirements/build.txt
# Default mount file as placeholder, this just avoid the mount error
ARG TORCH_WHEELS_PATH="./requirements"
# Install torch, torchaudio and torchvision
@ -332,13 +344,11 @@ RUN --mount=type=cache,target=/root/.cache/uv \
# Install xformers wheel from previous stage
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system /wheels/xformers/*.whl --verbose
# Build flashinfer from source.
ARG torch_cuda_arch_list='8.0;8.9;9.0a;10.0a;12.0'
# install package for build flashinfer
# see issue: https://github.com/flashinfer-ai/flashinfer/issues/738
RUN pip install build==1.3.0
RUN pip freeze | grep -E 'setuptools|packaging|build'
ENV TORCH_CUDA_ARCH_LIST=${torch_cuda_arch_list}

View File

@ -1,9 +1,14 @@
import glob
import os
requires_files = glob.glob("requirements/*.txt")
requires_files += ["pyproject.toml"]
for file in requires_files:
if not os.path.exists(file):
print(f"!!! skipping missing {file}")
continue
print(f">>> cleaning {file}")
with open(file) as f:
lines = f.readlines()

View File

@ -42,7 +42,7 @@ jobs:
build-external-packages: "vllm"
build-environment: linux-jammy-cuda12.8-py3.12-gcc11
docker-image-name: ci-image:pytorch-linux-jammy-cuda12.8-cudnn9-py3.12-gcc11-vllm
cuda-arch-list: '8.0;8.9;9.0'
cuda-arch-list: '8.0 8.9 9.0'
runner: linux.24xlarge.memory
test-matrix: |
{ include: [