# runner
**Repository Path**: ym2j/runner
## Basic Information
- **Project Name**: runner
- **Description**: No description available
- **Primary Language**: Unknown
- **License**: Apache-2.0
- **Default Branch**: main
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 0
- **Forks**: 0
- **Created**: 2025-11-25
- **Last Updated**: 2025-11-25
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
# GPUStack Runner
This repository serves as the Docker image pack center for GPUStack Runner.
It provides a collection of Dockerfiles to build images for various inference services across different accelerated
backends.
## Agenda
- [Onboard Services](#onboard-services)
- [Directory Structure](#directory-structure)
- [Dockerfile Convention](#dockerfile-convention)
- [Docker Image Naming Convention](#docker-image-naming-convention)
- [Integration Process](#integration-process)
## Onboard Services
> [!TIP]
> - The list below shows the accelerated backends and inference services available in the latest release. For support of
backends or services not shown here, please refer to previous release tags.
> - Deprecated inference service versions in the latest release are marked with strikethrough formatting. They may still
be available in previous releases, and not recommended for new deployments.
The following table lists the supported accelerated backends and their corresponding inference services with versions.
### Ascend CANN
| CANN Version (Variant) | Inference Service (Version List) |
|------------------------|------------------------------------------------------------------------------------------------------------------------------------------|
| 8.3 (A3/910C) | MindIE: `2.2.rc1`
vLLM: `0.11.0` |
| 8.3 (910B) | MindIE: `2.2.rc1`
vLLM: `0.11.0` |
| 8.3 (310P) | MindIE: `2.2.rc1` |
| 8.2 (A3/910C) | MindIE: `2.1.rc2`
vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`
SGLang: `0.5.2`, `0.5.1.post3` |
| 8.2 (910B) | MindIE: `2.1.rc2`, `2.1.rc1`
vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`, `0.9.2`, `0.9.1`
SGLang: `0.5.2`, `0.5.1.post3` |
| 8.2 (310P) | MindIE: `2.1.rc2`, `2.1.rc1`
vLLM: `0.10.0`, `0.9.2` |
### Iluvatar CoreX
| CoreX Version (Variant) | Inference Service (Version List) |
|-------------------------|----------------------------------|
| 4.2 | vLLM: `0.8.3` |
### NVIDIA CUDA
| CUDA Version (Variant) | Inference Service (Version List) |
|------------------------|-------------------------------------------------------------------------------------------------------------------------------------|
| 12.9 | vLLM: `0.11.2`
SGLang: `0.5.5.post3` |
| 12.8 | vLLM: `0.11.2`, `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
SGLang: `0.5.5.post3`, `0.5.5`, `0.5.4.post3`
VoxBox: `0.0.20` |
| 12.6 | vLLM: `0.11.2`, `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
SGLang: `0.5.5.post3`
VoxBox: `0.0.20` |
| 12.4 | vLLM: `0.11.0`, `0.10.2`, `0.10.1.1`, `0.10.0`
VoxBox: `0.0.20` |
> [!NOTE]
> - CUDA 12.8 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0 10.0+PTX 12.0+PTX`.
> - CUDA 12.6/12.4 supports Compute Capabilities: `7.5 8.0+PTX 8.9 9.0+PTX`.
### Hygon DTK
| DTK Version (Variant) | Inference Service (Version List) |
|-----------------------|----------------------------------|
| 25.04 | vLLM: `0.9.2`, `0.8.5` |
### MetaX MACA
| MACA Version (Variant) | Inference Service (Version List) |
|------------------------|----------------------------------|
| 3.2 | vLLM: `0.10.2` |
| 3.0 | vLLM: `0.9.1` |
### AMD ROCm
| ROCm Version (Variant) | Inference Service (Version List) |
|------------------------|------------------------------------------------------|
| 7.0 | vLLM: `0.11.2`, `0.11.0`
SGLang: `0.5.5.post3` |
| 6.4 | vLLM: `0.11.2`, `0.10.2`
SGLang: `0.5.5.post3` |
| 6.3 | vLLM: `0.10.1.1`, `0.10.0` |
> [!NOTE]
> - ROCm 7.0 supports LLVM Targets:
`gfx900 gfx906 gfx908 gfx90a gfx942 gfx1030 gfx1100 gfx1101 gfx1102 gfx1200 gfx1201`.
> - ROCm 6.4/6.3 supports LLVM Targets:
`gfx900 gfx906 gfx908 gfx90a gfx942 gfx1030 gfx1100 gfx1101 gfx1102 gfx1200 gfx1201`.
> - SGLang supports LLVM Targets: `gfx942`.
## Directory Structure
The pack skeleton is organized by backend:
```text
pack
├── {BACKEND 1}
│ └── Dockerfile
├── {BACKEND 2}
│ └── Dockerfile
├── {BACKEND 3}
│ └── Dockerfile
├── ...
│ └── Dockerfile
└── {BACKEND N}
└── Dockerfile
```
## Dockerfile Convention
Each Dockerfile follows these conventions:
- Begin with comments describing the package logic in steps and usage of build arguments (`ARG`s).
- Use `ARG` for all required and optional build arguments. If a required argument is unused, mark it as `(PLACEHOLDER)`.
- Use heredoc syntax for `RUN` commands to improve readability.
### Example Dockerfile Structure
```dockerfile
# Describe package logic and ARG usage.
#
ARG PYTHON_VERSION=... # REQUIRED
ARG CMAKE_MAX_JOBS=... # REQUIRED
ARG {OTHERS} # OPTIONAL
ARG {BACKEND}_VERSION=... # REQUIRED
ARG {BACKEND}_ARCHS=... # REQUIRED
ARG {BACKEND}_{OTHERS}=... # OPTIONAL
ARG {SERVICE}_BASE_IMAGE=... # REQUIRED
ARG {SERVICE}_VERSION=... # REQUIRED
ARG {SERVICE}_{OTHERS}=... # OPTIONAL
ARG {SERVICE}_{FRAMEWORK}_VERSION=... # REQUIRED
ARG {SERVICE}_{FRAMEWORK}_{OTHERS}=... # OPTIONAL
# Stage Bake Runtime
FROM {BACKEND DEVEL IMAGE} AS runtime
SHELL ["/bin/bash", "-eo", "pipefail", "-c"]
ARG TARGETPLATFORM
ARG TARGETOS
ARG TARGETARCH
ARG ...
RUN <