# runner **Repository Path**: ym2j/runner ## Basic Information - **Project Name**: runner - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-11-25 - **Last Updated**: 2025-11-25 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # GPUStack Runner This repository serves as the Docker image pack center for GPUStack Runner. It provides a collection of Dockerfiles to build images for various inference services across different accelerated backends. ## Agenda - [Onboard Services](#onboard-services) - [Directory Structure](#directory-structure) - [Dockerfile Convention](#dockerfile-convention) - [Docker Image Naming Convention](#docker-image-naming-convention) - [Integration Process](#integration-process) ## Onboard Services > [!TIP] > - The list below shows the accelerated backends and inference services available in the latest release. For support of backends or services not shown here, please refer to previous release tags. > - Deprecated inference service versions in the latest release are marked with strikethrough formatting. They may still be available in previous releases, and not recommended for new deployments. The following table lists the supported accelerated backends and their corresponding inference services with versions. ### Ascend CANN | CANN Version (Variant) | Inference Service (Version List) | |------------------------|------------------------------------------------------------------------------------------------------------------------------------------| | 8.3 (A3/910C) | MindIE: `2.2.rc1`
vLLM: `0.11.0` | | 8.3 (910B) | MindIE: `2.2.rc1`
vLLM: `0.11.0` | | 8.3 (310P) | MindIE: `2.2.rc1` | | 8.2 (A3/910C) | MindIE: `2.1.rc2`
vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`
SGLang: `0.5.2`, `0.5.1.post3` | | 8.2 (910B) | MindIE: `2.1.rc2`, `2.1.rc1`
vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`, `0.9.2`, `0.9.1`
SGLang: `0.5.2`, `0.5.1.post3` | | 8.2 (310P) | MindIE: `2.1.rc2`, `2.1.rc1`
vLLM: `0.10.0`, `0.9.2` | ### Iluvatar CoreX | CoreX Version (Variant) | Inference Service (Version List) | |-------------------------|----------------------------------| | 4.2 | vLLM: `0.8.3` | ### NVIDIA CUDA | CUDA Version (Variant) | Inference Service (Version List) | |------------------------|-------------------------------------------------------------------------------------------------------------------------------------| | 12.9 | vLLM: `0.11.2`
SGLang: `0.5.5.post3` | | 12.8 | vLLM: `0.11.2`, `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
SGLang: `0.5.5.post3`, `0.5.5`, `0.5.4.post3`
VoxBox: `0.0.20` | | 12.6 | vLLM: `0.11.2`, `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
SGLang: `0.5.5.post3`
VoxBox: `0.0.20` | | 12.4 | vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
VoxBox: `0.0.20` | > [!NOTE] > - CUDA 12.8 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0 10.0+PTX 12.0+PTX`. > - CUDA 12.6/12.4 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0+PTX`. ### Hygon DTK | DTK Version (Variant) | Inference Service (Version List) | |-----------------------|----------------------------------| | 25.04 | vLLM: `0.9.2`, `0.8.5` | ### MetaX MACA | MACA Version (Variant) | Inference Service (Version List) | |------------------------|----------------------------------| | 3.2 | vLLM: `0.10.2` | | 3.0 | vLLM: `0.9.1` | ### AMD ROCm | ROCm Version (Variant) | Inference Service (Version List) | |------------------------|------------------------------------------------------| | 7.0 | vLLM: `0.11.2`, `0.11.0`
SGLang: `0.5.5.post3` | | 6.4 | vLLM: `0.11.2`, `0.10.2`
SGLang: `0.5.5.post3` | | 6.3 | vLLM: `0.10.1.1`, `0.10.0` | > [!NOTE] > - ROCm 7.0 supports LLVM Targets: `gfx900 gfx906 gfx908 gfx90a gfx942 gfx1030 gfx1100 gfx1101 gfx1102 gfx1200 gfx1201`. > - ROCm 6.4/6.3 supports LLVM Targets: `gfx900 gfx906 gfx908 gfx90a gfx942 gfx1030 gfx1100 gfx1101 gfx1102 gfx1200 gfx1201`. > - SGLang supports LLVM Targets: `gfx942`. ## Directory Structure The pack skeleton is organized by backend: ```text pack ├── {BACKEND 1} │ └── Dockerfile ├── {BACKEND 2} │ └── Dockerfile ├── {BACKEND 3} │ └── Dockerfile ├── ... │ └── Dockerfile └── {BACKEND N} └── Dockerfile ``` ## Dockerfile Convention Each Dockerfile follows these conventions: - Begin with comments describing the package logic in steps and usage of build arguments (`ARG`s). - Use `ARG` for all required and optional build arguments. If a required argument is unused, mark it as `(PLACEHOLDER)`. - Use heredoc syntax for `RUN` commands to improve readability. ### Example Dockerfile Structure ```dockerfile # Describe package logic and ARG usage. # ARG PYTHON_VERSION=... # REQUIRED ARG CMAKE_MAX_JOBS=... # REQUIRED ARG {OTHERS} # OPTIONAL ARG {BACKEND}_VERSION=... # REQUIRED ARG {BACKEND}_ARCHS=... # REQUIRED ARG {BACKEND}_{OTHERS}=... # OPTIONAL ARG {SERVICE}_BASE_IMAGE=... # REQUIRED ARG {SERVICE}_VERSION=... # REQUIRED ARG {SERVICE}_{OTHERS}=... # OPTIONAL ARG {SERVICE}_{FRAMEWORK}_VERSION=... # REQUIRED ARG {SERVICE}_{FRAMEWORK}_{OTHERS}=... # OPTIONAL # Stage Bake Runtime FROM {BACKEND DEVEL IMAGE} AS runtime SHELL ["/bin/bash", "-eo", "pipefail", "-c"] ARG TARGETPLATFORM ARG TARGETOS ARG TARGETARCH ARG ... RUN <