| | |
| | |
| | FROM nvcr.io/nvidia/pytorch:24.08-py3 |
| |
|
| | |
| | ENV MAX_JOBS=32 |
| | ENV VLLM_WORKER_MULTIPROC_METHOD=spawn |
| | ENV DEBIAN_FRONTEND=noninteractive |
| | ENV NODE_OPTIONS="" |
| | ENV HF_HUB_ENABLE_HF_TRANSFER="1" |
| |
|
| | |
| | ARG APT_SOURCE=https://mirrors.tuna.tsinghua.edu.cn/ubuntu/ |
| | ARG PIP_INDEX=https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple |
| | ARG VLLM_COMMIT=227578480d71fc94ef46ca77fb69496412158d68 |
| |
|
| | |
| | RUN cp /etc/apt/sources.list /etc/apt/sources.list.bak && \ |
| | { \ |
| | echo "deb ${APT_SOURCE} jammy main restricted universe multiverse"; \ |
| | echo "deb ${APT_SOURCE} jammy-updates main restricted universe multiverse"; \ |
| | echo "deb ${APT_SOURCE} jammy-backports main restricted universe multiverse"; \ |
| | echo "deb ${APT_SOURCE} jammy-security main restricted universe multiverse"; \ |
| | } > /etc/apt/sources.list |
| |
|
| | |
| | RUN apt-get update && \ |
| | apt-get install -y -o Dpkg::Options::="--force-confdef" systemd && \ |
| | apt-get clean |
| |
|
| | |
| | RUN apt-get update && \ |
| | apt-get install -y tini && \ |
| | apt-get clean |
| |
|
| | |
| | RUN pip config set global.index-url "${PIP_INDEX}" && \ |
| | pip config set global.extra-index-url "${PIP_INDEX}" && \ |
| | python -m pip install --upgrade pip |
| |
|
| | |
| | RUN pip uninstall -y torch torchvision torchaudio \ |
| | pytorch-quantization pytorch-triton torch-tensorrt \ |
| | xgboost transformer_engine flash_attn apex megatron-core |
| |
|
| | |
| | RUN pip install --no-cache-dir vllm --pre --extra-index-url "https://wheels.vllm.ai/${VLLM_COMMIT}" && \ |
| | git clone -b verl_v1 https://github.com/hiyouga/vllm.git && \ |
| | cp -r vllm/vllm/ /usr/local/lib/python3.10/dist-packages/ |
| |
|
| | |
| | RUN pip install --no-cache-dir torch==2.5.1 torchvision==0.20.1 torchaudio==2.5.1 tensordict torchdata \ |
| | transformers>=4.49.0 accelerate datasets peft hf-transfer \ |
| | ray[default] codetiming hydra-core pandas pyarrow>=15.0.0 pylatexenc qwen-vl-utils wandb liger-kernel mathruler \ |
| | pytest yapf py-spy pyext pre-commit ruff |
| |
|
| | |
| | RUN wget -nv https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl && \ |
| | pip install --no-cache-dir flash_attn-2.7.4.post1+cu12torch2.5cxx11abiFALSE-cp310-cp310-linux_x86_64.whl |
| |
|
| | |
| | RUN pip uninstall -y pynvml nvidia-ml-py && \ |
| | pip install --no-cache-dir nvidia-ml-py>=12.560.30 opencv-python-headless==4.8.0.74 fastapi==0.115.6 && \ |
| | pip install --no-cache-dir --upgrade optree>=0.13.0 |
| |
|
| | |
| | RUN pip config unset global.index-url && \ |
| | pip config unset global.extra-index-url |
| |
|