https://arkprojects.space/wiki/AMD_GFX906
| Name | About | Status | Docs |
|---|---|---|---|
| ROCm | ROCm patched images | OK | readme |
| ROCm-tensile | gfx906 tensile files | OK | readme |
| PyTorch | PyTorch images | OK | readme |
| llama.cpp | llama.cpp images | OK | readme |
| ComfyUI | ComfyUI images | OK | readme |
| VLLM | vLLM images | OK | readme |
| Project | Image |
|---|---|
| ROCm | docker.io/mixa3607/rocm-gfx906:6.3.3-complete |
docker.io/mixa3607/rocm-gfx906:6.4.4-complete |
|
docker.io/mixa3607/rocm-gfx906:7.0.0-complete |
|
docker.io/mixa3607/rocm-gfx906:7.0.2-complete |
|
docker.io/mixa3607/rocm-gfx906:7.1.0-complete |
|
docker.io/mixa3607/rocm-gfx906:7.1.1-complete |
|
docker.io/mixa3607/rocm-gfx906:7.2.0-complete |
|
docker.io/mixa3607/rocm-gfx906:7.2.1-complete |
|
| --------- | ----------------------------------------------------------- |
| PyTorch | docker.io/mixa3607/pytorch-gfx906:v2.7.1-rocm-6.4.4 |
docker.io/mixa3607/pytorch-gfx906:v2.7.1-rocm-6.3.3 |
|
docker.io/mixa3607/pytorch-gfx906:v2.8.0-rocm-6.4.4 |
|
docker.io/mixa3607/pytorch-gfx906:v2.8.0-rocm-6.3.3 |
|
docker.io/mixa3607/pytorch-gfx906:v2.9.0-rocm-6.4.4 |
|
docker.io/mixa3607/pytorch-gfx906:v2.9.0-rocm-6.3.3 |
|
docker.io/mixa3607/pytorch-gfx906:v2.9.0-rocm-7.0.2 |
|
docker.io/mixa3607/pytorch-gfx906:v2.10.0-rocm-6.3.3 |
|
docker.io/mixa3607/pytorch-gfx906:v2.10.0-rocm-7.2.0 |
|
docker.io/mixa3607/pytorch-gfx906:v2.11.0-rocm-6.3.3 |
|
docker.io/mixa3607/pytorch-gfx906:v2.11.0-rocm-7.2.0 |
|
docker.io/mixa3607/pytorch-gfx906:v2.11.0-rocm-7.2.1 |
|
| --------- | ----------------------------------------------------------- |
| vLLM | docker.io/mixa3607/vllm-gfx906:0.20.1-rocm-6.3.3-aiinfos |
docker.io/mixa3607/vllm-gfx906:0.20.1-rocm-7.2.1-aiinfos |
|
| --------- | ----------------------------------------------------------- |
| ComfyUI | docker.io/mixa3607/comfyui-gfx906:<ver>-rocm-6.3.3* |
docker.io/mixa3607/comfyui-gfx906:<ver>-rocm-7.2.1* |
|
| --------- | ----------------------------------------------------------- |
| llama.cpp | docker.io/mixa3607/llama.cpp-gfx906:<ver>-rocm-6.3.3* |
docker.io/mixa3607/llama.cpp-gfx906:<ver>-rocm-7.2.1* |
* llama.cpp and ComfyUI have daily builds. See last tag on dockerhub
flowchart TD
rocm-src[docker.io/rocm/dev-ubuntu-24.04] --> rocm[docker.io/mixa3607/rocm-gfx906]
rocm --> llama[docker.io/mixa3607/llama.cpp-gfx906]
rocm --> torch[docker.io/mixa3607/pytorch-gfx906]
torch --> comfyui[docker.io/mixa3607/comfyui-gfx906]
torch --> vllm[docker.io/mixa3607/vllm-gfx906]
apt update && apt install -y rocm-validation-suite
echo 'actions:
- name: gst-581Tflops-4K4K8K-rand-bf16
device: all
module: gst
log_interval: 10000
ramp_interval: 5000
duration: 120000
hot_calls: 1000
copy_matrix: false
target_stress: 581000
matrix_size_a: 4864
matrix_size_b: 4096
matrix_size_c: 8192
matrix_init: rand
data_type: bf16_r
lda: 8320
ldb: 8320
ldc: 4992
ldd: 4992
transa: 1
transb: 0
alpha: 1
beta: 0' > ~/gst-581Tflops-4K4K8K-rand-bf16.conf
/opt/rocm/bin/rvs -c ~/gst-581Tflops-4K4K8K-rand-bf16.conf