From 579f0b73cbd69ecbb3b894d166e8523d2b95fb77 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Tue, 23 Sep 2025 13:39:28 -0400 Subject: [PATCH 01/20] Add build arg for ROCm version Signed-off-by: Simon Redman --- Dockerfile | 13 +++++++++++++ 1 file changed, 13 insertions(+) diff --git a/Dockerfile b/Dockerfile index a253237c974c..75a636b72d55 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,6 +19,7 @@ FROM requirements AS requirements-drivers ARG BUILD_TYPE ARG CUDA_MAJOR_VERSION=12 ARG CUDA_MINOR_VERSION=0 +ARG ROCM_VERSION=6.1.0 # ROCm version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `latest` or `7.0_alpha` or `6.3.4`. ARG SKIP_DRIVERS=false ARG TARGETARCH ARG TARGETVARIANT @@ -111,6 +112,18 @@ RUN if [ "${BUILD_TYPE}" = "clblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ ; fi RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ + # Setup for specific ROCm version as described here: https://rocm.docs.amd.com/projects/install-on-linux/en/latest/install/install-methods/package-manager/package-manager-ubuntu.html + apt-get update && \ + apt-get install -y --no-install-recommends \ + gpg wget && \ + mkdir --parents --mode=0755 /etc/apt/keyrings && \ + wget -qO - https://repo.radeon.com/rocm/rocm.gpg.key | gpg --yes --dearmor --output /etc/apt/keyrings/rocm.gpg && \ + echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/${ROCM_VERSION} jammy main" >> /etc/apt/sources.list.d/rocm.list && \ + echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/graphics/${ROCM_VERSION}/ubuntu jammy main" >> /etc/apt/sources.list.d/rocm.list && \ + echo "Package: *" >> /etc/apt/preferences.d/rocm-pin-600 && \ + echo "Pin: release o=repo.radeon.com" >> /etc/apt/preferences.d/rocm-pin-600 && \ + echo "Pin-Priority: 600" >> /etc/apt/preferences.d/rocm-pin-600 && \ + # End setup steps for specific ROCm version apt-get update && \ apt-get install -y --no-install-recommends \ hipblas-dev \ From 43522f88a4810d212569ed943dcad40486e2bbc1 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Tue, 23 Sep 2025 21:53:14 -0400 Subject: [PATCH 02/20] Break ROCM_VERSION into ROCM_{MAJOR,MINOR}_VERSION Signed-off-by: Simon Redman --- Dockerfile | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/Dockerfile b/Dockerfile index 75a636b72d55..b73083c1f937 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,7 +19,8 @@ FROM requirements AS requirements-drivers ARG BUILD_TYPE ARG CUDA_MAJOR_VERSION=12 ARG CUDA_MINOR_VERSION=0 -ARG ROCM_VERSION=6.1.0 # ROCm version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `latest` or `7.0_alpha` or `6.3.4`. +ARG ROCM_MAJOR_VERSION=6 +ARG ROCM_MINOR_VERSION=1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. ARG SKIP_DRIVERS=false ARG TARGETARCH ARG TARGETVARIANT @@ -113,13 +114,16 @@ RUN if [ "${BUILD_TYPE}" = "clblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ # Setup for specific ROCm version as described here: https://rocm.docs.amd.com/projects/install-on-linux/en/latest/install/install-methods/package-manager/package-manager-ubuntu.html + ROCM_VERSION="${ROCM_MAJOR_VERSION}.${ROCM_MINOR_VERSION}" && \ apt-get update && \ apt-get install -y --no-install-recommends \ gpg wget && \ mkdir --parents --mode=0755 /etc/apt/keyrings && \ wget -qO - https://repo.radeon.com/rocm/rocm.gpg.key | gpg --yes --dearmor --output /etc/apt/keyrings/rocm.gpg && \ echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/${ROCM_VERSION} jammy main" >> /etc/apt/sources.list.d/rocm.list && \ - echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/graphics/${ROCM_VERSION}/ubuntu jammy main" >> /etc/apt/sources.list.d/rocm.list && \ + if [ "${ROCM_MAJOR_VERSION}" -ge 7 ]; then \ + echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/graphics/${ROCM_VERSION}/ubuntu jammy main" >> /etc/apt/sources.list.d/rocm.list \ + ; fi && \ echo "Package: *" >> /etc/apt/preferences.d/rocm-pin-600 && \ echo "Pin: release o=repo.radeon.com" >> /etc/apt/preferences.d/rocm-pin-600 && \ echo "Pin-Priority: 600" >> /etc/apt/preferences.d/rocm-pin-600 && \ From 3297f64c530cf11b5fdc422b19b9c3c8813d3b7e Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Tue, 23 Sep 2025 21:53:26 -0400 Subject: [PATCH 03/20] Use correct ROCm package names Signed-off-by: Simon Redman --- Dockerfile | 3 +++ 1 file changed, 3 insertions(+) diff --git a/Dockerfile b/Dockerfile index b73083c1f937..29268d4d815f 100644 --- a/Dockerfile +++ b/Dockerfile @@ -130,6 +130,9 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then # End setup steps for specific ROCm version apt-get update && \ apt-get install -y --no-install-recommends \ + rocm-developer-tools \ + rocm-hip-runtime-dev \ + rocm-hip-sdk \ hipblas-dev \ rocblas-dev && \ apt-get clean && \ From 2b38e07d880caaad855fb2e734af5ec5997c8e8a Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Thu, 25 Sep 2025 18:16:18 -0400 Subject: [PATCH 04/20] Add rocm package for runtime libs Signed-off-by: Simon Redman --- Dockerfile | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index 29268d4d815f..be11c8b6d41a 100644 --- a/Dockerfile +++ b/Dockerfile @@ -130,11 +130,14 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then # End setup steps for specific ROCm version apt-get update && \ apt-get install -y --no-install-recommends \ + # Build dependencies rocm-developer-tools \ rocm-hip-runtime-dev \ rocm-hip-sdk \ hipblas-dev \ - rocblas-dev && \ + rocblas-dev \ + # Metapackage for the ROCm runtime + client tools + rocm && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ echo "amd" > /run/localai/capability && \ From 8a2451f3bbe35a7863cdda027859be1ae98f4850 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Thu, 25 Sep 2025 18:16:52 -0400 Subject: [PATCH 05/20] Remove hipblas-dev and rocblas-dev. I think they are not needed, and have been replaced by the libraries from AMD Signed-off-by: Simon Redman --- Dockerfile | 2 -- 1 file changed, 2 deletions(-) diff --git a/Dockerfile b/Dockerfile index be11c8b6d41a..18acc03b038e 100644 --- a/Dockerfile +++ b/Dockerfile @@ -134,8 +134,6 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then rocm-developer-tools \ rocm-hip-runtime-dev \ rocm-hip-sdk \ - hipblas-dev \ - rocblas-dev \ # Metapackage for the ROCm runtime + client tools rocm && \ apt-get clean && \ From bf5ccc3fd80d52752a8f451da9ac0dba02a8af33 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Tue, 16 Dec 2025 14:12:04 -0500 Subject: [PATCH 06/20] Migrate ROCM build flags to llama-cpp docker file Signed-off-by: Simon Redman --- backend/Dockerfile.llama-cpp | 25 +++++++++++++++++++++++-- 1 file changed, 23 insertions(+), 2 deletions(-) diff --git a/backend/Dockerfile.llama-cpp b/backend/Dockerfile.llama-cpp index 8d5e0b7fbdb0..e5a75147e5f1 100644 --- a/backend/Dockerfile.llama-cpp +++ b/backend/Dockerfile.llama-cpp @@ -58,6 +58,8 @@ ARG CUDA_MINOR_VERSION ARG SKIP_DRIVERS=false ENV CUDA_MAJOR_VERSION=${CUDA_MAJOR_VERSION} ENV CUDA_MINOR_VERSION=${CUDA_MINOR_VERSION} +ARG ROCM_MAJOR_VERSION=7 +ARG ROCM_MINOR_VERSION=1.1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. ENV DEBIAN_FRONTEND=noninteractive ARG TARGETARCH ARG TARGETVARIANT @@ -158,10 +160,29 @@ RUN if [ "${BUILD_TYPE}" = "clblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ ; fi RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \ + # Setup for specific ROCm version as described here: https://rocm.docs.amd.com/projects/install-on-linux/en/latest/install/install-methods/package-manager/package-manager-ubuntu.html + ROCM_VERSION="${ROCM_MAJOR_VERSION}.${ROCM_MINOR_VERSION}" && \ apt-get update && \ apt-get install -y --no-install-recommends \ - hipblas-dev \ - rocblas-dev && \ + gpg wget && \ + mkdir --parents --mode=0755 /etc/apt/keyrings && \ + wget -qO - https://repo.radeon.com/rocm/rocm.gpg.key | gpg --yes --dearmor --output /etc/apt/keyrings/rocm.gpg && \ + echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/${ROCM_VERSION} jammy main" >> /etc/apt/sources.list.d/rocm.list && \ + if [ "${ROCM_MAJOR_VERSION}" -ge 7 ]; then \ + echo "deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/graphics/${ROCM_VERSION}/ubuntu jammy main" >> /etc/apt/sources.list.d/rocm.list \ + ; fi && \ + echo "Package: *" >> /etc/apt/preferences.d/rocm-pin-600 && \ + echo "Pin: release o=repo.radeon.com" >> /etc/apt/preferences.d/rocm-pin-600 && \ + echo "Pin-Priority: 600" >> /etc/apt/preferences.d/rocm-pin-600 && \ + # End setup steps for specific ROCm version + apt-get update && \ + apt-get install -y --no-install-recommends \ + # Build dependencies + rocm-developer-tools \ + rocm-hip-runtime-dev \ + rocm-hip-sdk \ + # Metapackage for the ROCm runtime + client tools + rocm && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ # I have no idea why, but the ROCM lib packages don't trigger ldconfig after they install, which results in local-ai and others not being able From 9cbe3a6143244a457e9d248b1c185ddf539a4cbb Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Wed, 17 Dec 2025 22:51:42 -0500 Subject: [PATCH 07/20] Change base-level Dockerfile back to same runtime dependency packages as previously needed Signed-off-by: Simon Redman --- Dockerfile | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/Dockerfile b/Dockerfile index 18acc03b038e..311f436ab79c 100644 --- a/Dockerfile +++ b/Dockerfile @@ -127,15 +127,11 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then echo "Package: *" >> /etc/apt/preferences.d/rocm-pin-600 && \ echo "Pin: release o=repo.radeon.com" >> /etc/apt/preferences.d/rocm-pin-600 && \ echo "Pin-Priority: 600" >> /etc/apt/preferences.d/rocm-pin-600 && \ - # End setup steps for specific ROCm version + # End setup steps for specific ROCm version - the packages below will be installed from the configured repositories apt-get update && \ apt-get install -y --no-install-recommends \ - # Build dependencies - rocm-developer-tools \ - rocm-hip-runtime-dev \ - rocm-hip-sdk \ - # Metapackage for the ROCm runtime + client tools - rocm && \ + rocblas-dev \ + hipblas-dev && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ echo "amd" > /run/localai/capability && \ From 64a111e0f754c77cbcbb804c0953502c2fbb2896 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Wed, 17 Dec 2025 22:52:03 -0500 Subject: [PATCH 08/20] Use default ROCm version of 5.5.1 to match Ubuntu package Signed-off-by: Simon Redman --- Dockerfile | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/Dockerfile b/Dockerfile index 311f436ab79c..a60fc8915ea1 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,8 +19,8 @@ FROM requirements AS requirements-drivers ARG BUILD_TYPE ARG CUDA_MAJOR_VERSION=12 ARG CUDA_MINOR_VERSION=0 -ARG ROCM_MAJOR_VERSION=6 -ARG ROCM_MINOR_VERSION=1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. +ARG ROCM_MAJOR_VERSION=5 +ARG ROCM_MINOR_VERSION=5.1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. ARG SKIP_DRIVERS=false ARG TARGETARCH ARG TARGETVARIANT From 958fb9d883cd62cf0412930bd26d092c3699f7ea Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Wed, 17 Dec 2025 22:55:14 -0500 Subject: [PATCH 09/20] Remove ROCm runtime package from llamacpp dockerfile Signed-off-by: Simon Redman --- backend/Dockerfile.llama-cpp | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/backend/Dockerfile.llama-cpp b/backend/Dockerfile.llama-cpp index e5a75147e5f1..f3221022e0f7 100644 --- a/backend/Dockerfile.llama-cpp +++ b/backend/Dockerfile.llama-cpp @@ -55,11 +55,13 @@ ARG BUILD_TYPE ENV BUILD_TYPE=${BUILD_TYPE} ARG CUDA_MAJOR_VERSION ARG CUDA_MINOR_VERSION +ARG ROCM_MAJOR_VERSION +ARG ROCM_MINOR_VERSION ARG SKIP_DRIVERS=false ENV CUDA_MAJOR_VERSION=${CUDA_MAJOR_VERSION} ENV CUDA_MINOR_VERSION=${CUDA_MINOR_VERSION} -ARG ROCM_MAJOR_VERSION=7 -ARG ROCM_MINOR_VERSION=1.1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. +ENV ROCM_MAJOR_VERSION=${ROCM_MAJOR_VERSION} +ENV ROCM_MINOR_VERSION=${ROCM_MINOR_VERSION} ENV DEBIAN_FRONTEND=noninteractive ARG TARGETARCH ARG TARGETVARIANT @@ -180,9 +182,7 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then # Build dependencies rocm-developer-tools \ rocm-hip-runtime-dev \ - rocm-hip-sdk \ - # Metapackage for the ROCm runtime + client tools - rocm && \ + rocm-hip-sdk && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ # I have no idea why, but the ROCM lib packages don't trigger ldconfig after they install, which results in local-ai and others not being able From d407465cbdef2da06b22c9a9c7cadaad49bc2c37 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Thu, 18 Dec 2025 10:08:34 -0500 Subject: [PATCH 10/20] Annotate ROCm runtime dependencies Signed-off-by: Simon Redman --- Dockerfile | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/Dockerfile b/Dockerfile index a60fc8915ea1..61a867774fc4 100644 --- a/Dockerfile +++ b/Dockerfile @@ -130,8 +130,9 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then # End setup steps for specific ROCm version - the packages below will be installed from the configured repositories apt-get update && \ apt-get install -y --no-install-recommends \ - rocblas-dev \ - hipblas-dev && \ + rocm-hip-runtime # Gives libomp.so\ + rocblas-dev # Give librocblas.so \ + hipblas-dev # Gives libhipblas.so && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ echo "amd" > /run/localai/capability && \ From d562299b178f75dae31d7fc92c35e2832c57cac7 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Thu, 18 Dec 2025 20:14:31 -0500 Subject: [PATCH 11/20] Remove comments in runtime dependencies install step Signed-off-by: Simon Redman --- Dockerfile | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/Dockerfile b/Dockerfile index 61a867774fc4..ab555eb89fbc 100644 --- a/Dockerfile +++ b/Dockerfile @@ -130,9 +130,9 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then # End setup steps for specific ROCm version - the packages below will be installed from the configured repositories apt-get update && \ apt-get install -y --no-install-recommends \ - rocm-hip-runtime # Gives libomp.so\ - rocblas-dev # Give librocblas.so \ - hipblas-dev # Gives libhipblas.so && \ + rocm-hip-runtime \ + rocblas-dev \ + hipblas-dev && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ echo "amd" > /run/localai/capability && \ From 2542449809c0850de9b28001a06541b47d85b6bb Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Fri, 19 Dec 2025 20:21:43 -0500 Subject: [PATCH 12/20] Change default ROCM version to 6.4.3 Signed-off-by: Simon Redman --- Dockerfile | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/Dockerfile b/Dockerfile index ab555eb89fbc..4e289f78b6af 100644 --- a/Dockerfile +++ b/Dockerfile @@ -19,8 +19,8 @@ FROM requirements AS requirements-drivers ARG BUILD_TYPE ARG CUDA_MAJOR_VERSION=12 ARG CUDA_MINOR_VERSION=0 -ARG ROCM_MAJOR_VERSION=5 -ARG ROCM_MINOR_VERSION=5.1 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. +ARG ROCM_MAJOR_VERSION=6 +ARG ROCM_MINOR_VERSION=4.3 # ROCm version to append to the major version, in the format of their apt repo (https://repo.radeon.com/rocm/apt/). Like `0_alpha` or `3.4`. ARG SKIP_DRIVERS=false ARG TARGETARCH ARG TARGETVARIANT From 8f1036cf160b2c2f34f587a9c7926e817db5b59c Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Fri, 19 Dec 2025 20:31:09 -0500 Subject: [PATCH 13/20] Rename ROCm6 build in image CI output to -gpu-amd-rocm-6 Signed-off-by: Simon Redman --- .github/workflows/backend.yml | 30 +++++++++++++++--------------- .github/workflows/image.yml | 9 +++++---- 2 files changed, 20 insertions(+), 19 deletions(-) diff --git a/.github/workflows/backend.yml b/.github/workflows/backend.yml index 66b1c683b88f..f9557758f7e2 100644 --- a/.github/workflows/backend.yml +++ b/.github/workflows/backend.yml @@ -634,7 +634,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-rerankers' + tag-suffix: '-gpu-amd-rocm-6-rerankers' runs-on: 'ubuntu-latest' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -643,13 +643,13 @@ jobs: context: "./backend" ubuntu-version: '2204' - build-type: 'hipblas' - cuda-major-version: "" - cuda-minor-version: "" + rocm-major-version: "6" + rocm-minor-version: "4.3" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-llama-cpp' + tag-suffix: '-gpu-amd-rocm-6-llama-cpp' runs-on: 'ubuntu-latest' - base-image: "rocm/dev-ubuntu-22.04:6.4.3" + base-image: "ubuntu:22.04" skip-drivers: 'false' backend: "llama-cpp" dockerfile: "./backend/Dockerfile.llama-cpp" @@ -660,7 +660,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-vllm' + tag-suffix: '-gpu-amd-rocm-6-vllm' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -673,7 +673,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-transformers' + tag-suffix: '-gpu-amd-rocm-6-transformers' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -686,7 +686,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-diffusers' + tag-suffix: '-gpu-amd-rocm-6-diffusers' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -700,7 +700,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-kokoro' + tag-suffix: '-gpu-amd-rocm-6-kokoro' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -713,7 +713,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-vibevoice' + tag-suffix: '-gpu-amd-rocm-6-vibevoice' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -726,7 +726,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-faster-whisper' + tag-suffix: '-gpu-amd-rocm-6-faster-whisper' runs-on: 'ubuntu-latest' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -739,7 +739,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-coqui' + tag-suffix: '-gpu-amd-rocm-6-coqui' runs-on: 'ubuntu-latest' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -752,7 +752,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-bark' + tag-suffix: '-gpu-amd-rocm-6-bark' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' @@ -1161,7 +1161,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-whisper' + tag-suffix: '-gpu-amd-rocm-6-whisper' base-image: "rocm/dev-ubuntu-22.04:6.4.3" runs-on: 'ubuntu-latest' skip-drivers: 'false' @@ -1376,7 +1376,7 @@ jobs: cuda-minor-version: "" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-rocm-hipblas-neutts' + tag-suffix: '-gpu-amd-rocm-6-neutts' runs-on: 'arc-runner-set' base-image: "rocm/dev-ubuntu-22.04:6.4.3" skip-drivers: 'false' diff --git a/.github/workflows/image.yml b/.github/workflows/image.yml index 7389760912c5..ec5daaf40864 100644 --- a/.github/workflows/image.yml +++ b/.github/workflows/image.yml @@ -37,14 +37,15 @@ jobs: matrix: include: - build-type: 'hipblas' + rocm-major-version: "6" + rocm-minor-version: "4.3" platforms: 'linux/amd64' tag-latest: 'auto' - tag-suffix: '-gpu-hipblas' - base-image: "rocm/dev-ubuntu-22.04:6.4.3" - grpc-base-image: "ubuntu:22.04" + tag-suffix: '-gpu-amd-rocm-6' + base-image: "ubuntu:22.04" runs-on: 'ubuntu-latest' makeflags: "--jobs=3 --output-sync=target" - aio: "-aio-gpu-hipblas" + aio: "-aio-gpu-amd-rocm-6" ubuntu-version: '2204' core-image-build: From e8f17b73830ca0a83c5e26d6702c2e3c873536db Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Fri, 19 Dec 2025 20:32:13 -0500 Subject: [PATCH 14/20] Add ROCm 7 image builds Signed-off-by: Simon Redman --- .github/workflows/backend.yml | 13 +++++++++++++ .github/workflows/image.yml | 11 +++++++++++ 2 files changed, 24 insertions(+) diff --git a/.github/workflows/backend.yml b/.github/workflows/backend.yml index f9557758f7e2..7534f537da90 100644 --- a/.github/workflows/backend.yml +++ b/.github/workflows/backend.yml @@ -655,6 +655,19 @@ jobs: dockerfile: "./backend/Dockerfile.llama-cpp" context: "./" ubuntu-version: '2204' + - build-type: 'hipblas' + rocm-major-version: "7" + rocm-minor-version: "1.1" + platforms: 'linux/amd64' + tag-latest: 'auto' + tag-suffix: '-gpu-amd-rocm-7-llama-cpp' + runs-on: 'ubuntu-latest' + base-image: "ubuntu:22.04" + skip-drivers: 'false' + backend: "llama-cpp" + dockerfile: "./backend/Dockerfile.llama-cpp" + context: "./" + ubuntu-version: '2204' - build-type: 'hipblas' cuda-major-version: "" cuda-minor-version: "" diff --git a/.github/workflows/image.yml b/.github/workflows/image.yml index ec5daaf40864..027eec702129 100644 --- a/.github/workflows/image.yml +++ b/.github/workflows/image.yml @@ -47,6 +47,17 @@ jobs: makeflags: "--jobs=3 --output-sync=target" aio: "-aio-gpu-amd-rocm-6" ubuntu-version: '2204' + - build-type: 'hipblas' + rocm-major-version: "7" + rocm-minor-version: "1.1" + platforms: 'linux/amd64' + tag-latest: 'auto' + tag-suffix: '-gpu-amd-rocm-7' + base-image: "ubuntu:22.04" + runs-on: 'ubuntu-latest' + makeflags: "--jobs=3 --output-sync=target" + aio: "-aio-gpu-amd-rocm-7" + ubuntu-version: '2204' core-image-build: uses: ./.github/workflows/image_build.yml From 748352c18e84f3bb0e02e1d26d4c322faff8052f Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:03:01 -0500 Subject: [PATCH 15/20] Add rocm-*-version arguments to backend.yml header Signed-off-by: Simon Redman --- .github/workflows/backend.yml | 2 ++ 1 file changed, 2 insertions(+) diff --git a/.github/workflows/backend.yml b/.github/workflows/backend.yml index 7534f537da90..6942c9b32cbb 100644 --- a/.github/workflows/backend.yml +++ b/.github/workflows/backend.yml @@ -21,6 +21,8 @@ jobs: build-type: ${{ matrix.build-type }} cuda-major-version: ${{ matrix.cuda-major-version }} cuda-minor-version: ${{ matrix.cuda-minor-version }} + rocm-major-version: ${{ matrix.rocm-major-version }} + rocm-minor-version: ${{ matrix.rocm-minor-version }} platforms: ${{ matrix.platforms }} runs-on: ${{ matrix.runs-on }} base-image: ${{ matrix.base-image }} From b0eb4b37a92c9c5d9fc8268f40bde53bf7d303b6 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:03:23 -0500 Subject: [PATCH 16/20] Change 'amd' capability to 'amd-rocm-' Signed-off-by: Simon Redman --- Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index 4e289f78b6af..742c38e3c677 100644 --- a/Dockerfile +++ b/Dockerfile @@ -135,7 +135,7 @@ RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then hipblas-dev && \ apt-get clean && \ rm -rf /var/lib/apt/lists/* && \ - echo "amd" > /run/localai/capability && \ + echo "amd-rocm-${ROCM_MAJOR_VERSION}" > /run/localai/capability && \ # I have no idea why, but the ROCM lib packages don't trigger ldconfig after they install, which results in local-ai and others not being able # to locate the libraries. We run ldconfig ourselves to work around this packaging deficiency ldconfig \ From bc9eafa155e357038799bfd9fde760182c9afc25 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:07:08 -0500 Subject: [PATCH 17/20] Translate all backend index.yaml entries to use 'amd-rocm-6' instead of 'amd' Signed-off-by: Simon Redman --- backend/index.yaml | 192 ++++++++++++++++++++++----------------------- 1 file changed, 96 insertions(+), 96 deletions(-) diff --git a/backend/index.yaml b/backend/index.yaml index a383a15fdf3d..6042e9f8b6fc 100644 --- a/backend/index.yaml +++ b/backend/index.yaml @@ -21,7 +21,7 @@ default: "cpu-llama-cpp" nvidia: "cuda12-llama-cpp" intel: "intel-sycl-f16-llama-cpp" - amd: "rocm-llama-cpp" + amd-rocm-6: "rocm6-llama-cpp" metal: "metal-llama-cpp" vulkan: "vulkan-llama-cpp" nvidia-l4t: "nvidia-l4t-arm64-llama-cpp" @@ -49,7 +49,7 @@ nvidia: "cuda12-whisper" intel: "intel-sycl-f16-whisper" metal: "metal-whisper" - amd: "rocm-whisper" + amd-rocm-6: "rocm6-whisper" vulkan: "vulkan-whisper" nvidia-l4t: "nvidia-l4t-arm64-whisper" nvidia-cuda-13: "cuda13-whisper" @@ -76,7 +76,7 @@ default: "cpu-stablediffusion-ggml" nvidia: "cuda12-stablediffusion-ggml" intel: "intel-sycl-f16-stablediffusion-ggml" - # amd: "rocm-stablediffusion-ggml" + # amd-rocm-6: "rocm6-stablediffusion-ggml" vulkan: "vulkan-stablediffusion-ggml" nvidia-l4t: "nvidia-l4t-arm64-stablediffusion-ggml" metal: "metal-stablediffusion-ggml" @@ -103,7 +103,7 @@ capabilities: nvidia: "cuda12-rfdetr" intel: "intel-rfdetr" - #amd: "rocm-rfdetr" + #amd-rocm-6: "rocm6-rfdetr" nvidia-l4t: "nvidia-l4t-arm64-rfdetr" default: "cpu-rfdetr" nvidia-cuda-13: "cuda13-rfdetr" @@ -139,7 +139,7 @@ alias: "vllm" capabilities: nvidia: "cuda12-vllm" - amd: "rocm-vllm" + amd-rocm-6: "rocm6-vllm" intel: "intel-vllm" nvidia-cuda-12: "cuda12-vllm" - &mlx @@ -197,7 +197,7 @@ capabilities: nvidia: "cuda12-rerankers" intel: "intel-rerankers" - amd: "rocm-rerankers" + amd-rocm-6: "rocm6-rerankers" - &transformers name: "transformers" icon: https://camo.githubusercontent.com/26569a27b8a30a488dd345024b71dbc05da7ff1b2ba97bb6080c9f1ee0f26cc7/68747470733a2f2f68756767696e67666163652e636f2f64617461736574732f68756767696e67666163652f646f63756d656e746174696f6e2d696d616765732f7265736f6c76652f6d61696e2f7472616e73666f726d6572732f7472616e73666f726d6572735f61735f615f6d6f64656c5f646566696e6974696f6e2e706e67 @@ -214,7 +214,7 @@ capabilities: nvidia: "cuda12-transformers" intel: "intel-transformers" - amd: "rocm-transformers" + amd-rocm-6: "rocm6-transformers" nvidia-cuda-13: "cuda13-transformers" nvidia-cuda-12: "cuda12-transformers" - &diffusers @@ -233,7 +233,7 @@ capabilities: nvidia: "cuda12-diffusers" intel: "intel-diffusers" - amd: "rocm-diffusers" + amd-rocm-6: "rocm6-diffusers" nvidia-l4t: "nvidia-l4t-diffusers" metal: "metal-diffusers" default: "cpu-diffusers" @@ -272,7 +272,7 @@ capabilities: nvidia: "cuda12-faster-whisper" intel: "intel-faster-whisper" - amd: "rocm-faster-whisper" + amd-rocm-6: "rocm6-faster-whisper" nvidia-cuda-13: "cuda13-faster-whisper" nvidia-cuda-12: "cuda12-faster-whisper" - &kokoro @@ -292,7 +292,7 @@ capabilities: nvidia: "cuda12-kokoro" intel: "intel-kokoro" - amd: "rocm-kokoro" + amd-rocm-6: "rocm6-kokoro" nvidia-l4t: "nvidia-l4t-kokoro" nvidia-cuda-13: "cuda13-kokoro" nvidia-cuda-12: "cuda12-kokoro" @@ -317,7 +317,7 @@ capabilities: nvidia: "cuda12-coqui" intel: "intel-coqui" - amd: "rocm-coqui" + amd-rocm-6: "rocm6-coqui" nvidia-cuda-13: "cuda13-coqui" nvidia-cuda-12: "cuda12-coqui" icon: https://avatars.githubusercontent.com/u/1338804?s=200&v=4 @@ -402,7 +402,7 @@ capabilities: nvidia: "cuda12-vibevoice" intel: "intel-vibevoice" - amd: "rocm-vibevoice" + amd-rocm-6: "rocm6-vibevoice" nvidia-l4t: "nvidia-l4t-vibevoice" default: "cpu-vibevoice" nvidia-cuda-13: "cuda13-vibevoice" @@ -496,7 +496,7 @@ capabilities: default: "cpu-neutts" nvidia: "cuda12-neutts" - amd: "rocm-neutts" + amd-rocm-6: "rocm6-neutts" nvidia-l4t: "nvidia-l4t-neutts" nvidia-cuda-12: "cuda12-neutts" nvidia-l4t-cuda-12: "nvidia-l4t-arm64-neutts" @@ -505,7 +505,7 @@ capabilities: default: "cpu-neutts-development" nvidia: "cuda12-neutts-development" - amd: "rocm-neutts-development" + amd-rocm-6: "rocm6-neutts-development" nvidia-l4t: "nvidia-l4t-neutts-development" nvidia-cuda-12: "cuda12-neutts-development" nvidia-l4t-cuda-12: "nvidia-l4t-arm64-neutts-development" @@ -515,7 +515,7 @@ default: "cpu-llama-cpp-development" nvidia: "cuda12-llama-cpp-development" intel: "intel-sycl-f16-llama-cpp-development" - amd: "rocm-llama-cpp-development" + amd-rocm-6: "rocm6-llama-cpp-development" metal: "metal-llama-cpp-development" vulkan: "vulkan-llama-cpp-development" nvidia-l4t: "nvidia-l4t-arm64-llama-cpp-development" @@ -534,10 +534,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-neutts - !!merge <<: *neutts - name: "rocm-neutts" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-neutts" + name: "rocm6-neutts" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-neutts" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-neutts + - localai/localai-backends:latest-gpu-amd-rocm-6-neutts - !!merge <<: *neutts name: "nvidia-l4t-arm64-neutts" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-neutts" @@ -554,10 +554,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-neutts - !!merge <<: *neutts - name: "rocm-neutts-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-neutts" + name: "rocm6-neutts-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-neutts" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-neutts + - localai/localai-backends:master-gpu-amd-rocm-6-neutts - !!merge <<: *neutts name: "nvidia-l4t-arm64-neutts-development" uri: "quay.io/go-skynet/local-ai-backends:master-nvidia-l4t-arm64-neutts" @@ -645,10 +645,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-llama-cpp - !!merge <<: *llamacpp - name: "rocm-llama-cpp" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-llama-cpp" + name: "rocm6-llama-cpp" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-llama-cpp" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-llama-cpp + - localai/localai-backends:latest-gpu-amd-rocm-6-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-llama-cpp" @@ -690,10 +690,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-llama-cpp - !!merge <<: *llamacpp - name: "rocm-llama-cpp-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-llama-cpp" + name: "rocm6-llama-cpp-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-llama-cpp" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-llama-cpp + - localai/localai-backends:master-gpu-amd-rocm-6-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-llama-cpp" @@ -766,10 +766,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-whisper - !!merge <<: *whispercpp - name: "rocm-whisper" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-whisper" + name: "rocm6-whisper" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-whisper" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-whisper + - localai/localai-backends:latest-gpu-amd-rocm-6-whisper - !!merge <<: *whispercpp name: "intel-sycl-f32-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-whisper" @@ -811,10 +811,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-whisper - !!merge <<: *whispercpp - name: "rocm-whisper-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-whisper" + name: "rocm6-whisper-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-whisper" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-whisper + - localai/localai-backends:master-gpu-amd-rocm-6-whisper - !!merge <<: *whispercpp name: "intel-sycl-f32-whisper-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-whisper" @@ -939,7 +939,7 @@ name: "vllm-development" capabilities: nvidia: "cuda12-vllm-development" - amd: "rocm-vllm-development" + amd-rocm-6: "rocm6-vllm-development" intel: "intel-vllm-development" - !!merge <<: *vllm name: "cuda12-vllm" @@ -947,10 +947,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-vllm - !!merge <<: *vllm - name: "rocm-vllm" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-vllm" + name: "rocm6-vllm" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-vllm" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-vllm + - localai/localai-backends:latest-gpu-amd-rocm-6-vllm - !!merge <<: *vllm name: "intel-vllm" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-vllm" @@ -962,10 +962,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-vllm - !!merge <<: *vllm - name: "rocm-vllm-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-vllm" + name: "rocm6-vllm-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-vllm" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-vllm + - localai/localai-backends:master-gpu-amd-rocm-6-vllm - !!merge <<: *vllm name: "intel-vllm-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-vllm" @@ -977,7 +977,7 @@ capabilities: nvidia: "cuda12-rfdetr-development" intel: "intel-rfdetr-development" - #amd: "rocm-rfdetr-development" + #amd-rocm-6: "rocm6-rfdetr-development" nvidia-l4t: "nvidia-l4t-arm64-rfdetr-development" default: "cpu-rfdetr-development" nvidia-cuda-13: "cuda13-rfdetr-development" @@ -992,7 +992,7 @@ mirrors: - localai/localai-backends:latest-gpu-intel-rfdetr # - !!merge <<: *rfdetr -# name: "rocm-rfdetr" +# name: "rocm6-rfdetr" # uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-hipblas-rfdetr" # mirrors: # - localai/localai-backends:latest-gpu-hipblas-rfdetr @@ -1022,7 +1022,7 @@ mirrors: - localai/localai-backends:master-gpu-intel-rfdetr # - !!merge <<: *rfdetr -# name: "rocm-rfdetr-development" +# name: "rocm6-rfdetr-development" # uri: "quay.io/go-skynet/local-ai-backends:master-gpu-hipblas-rfdetr" # mirrors: # - localai/localai-backends:master-gpu-hipblas-rfdetr @@ -1052,7 +1052,7 @@ capabilities: nvidia: "cuda12-rerankers-development" intel: "intel-rerankers-development" - amd: "rocm-rerankers-development" + amd-rocm-6: "rocm6-rerankers-development" nvidia-cuda-13: "cuda13-rerankers-development" - !!merge <<: *rerankers name: "cuda11-rerankers" @@ -1070,10 +1070,10 @@ mirrors: - localai/localai-backends:latest-gpu-intel-rerankers - !!merge <<: *rerankers - name: "rocm-rerankers" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-rerankers" + name: "rocm6-rerankers" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-rerankers" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-rerankers + - localai/localai-backends:latest-gpu-amd-rocm-6-rerankers - !!merge <<: *rerankers name: "cuda11-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-rerankers" @@ -1085,10 +1085,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-rerankers - !!merge <<: *rerankers - name: "rocm-rerankers-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-rerankers" + name: "rocm6-rerankers-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-rerankers" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-rerankers + - localai/localai-backends:master-gpu-amd-rocm-6-rerankers - !!merge <<: *rerankers name: "intel-rerankers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-rerankers" @@ -1110,7 +1110,7 @@ capabilities: nvidia: "cuda12-transformers-development" intel: "intel-transformers-development" - amd: "rocm-transformers-development" + amd-rocm-6: "rocm6-transformers-development" nvidia-cuda-13: "cuda13-transformers-development" - !!merge <<: *transformers name: "cuda12-transformers" @@ -1118,10 +1118,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-transformers - !!merge <<: *transformers - name: "rocm-transformers" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-transformers" + name: "rocm6-transformers" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-transformers" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-transformers + - localai/localai-backends:latest-gpu-amd-rocm-6-transformers - !!merge <<: *transformers name: "intel-transformers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-transformers" @@ -1143,10 +1143,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-transformers - !!merge <<: *transformers - name: "rocm-transformers-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-transformers" + name: "rocm6-transformers-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-transformers" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-transformers + - localai/localai-backends:master-gpu-amd-rocm-6-transformers - !!merge <<: *transformers name: "intel-transformers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-transformers" @@ -1168,7 +1168,7 @@ capabilities: nvidia: "cuda12-diffusers-development" intel: "intel-diffusers-development" - amd: "rocm-diffusers-development" + amd-rocm-6: "rocm6-diffusers-development" nvidia-l4t: "nvidia-l4t-diffusers-development" metal: "metal-diffusers-development" default: "cpu-diffusers-development" @@ -1209,10 +1209,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-diffusers - !!merge <<: *diffusers - name: "rocm-diffusers" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-diffusers" + name: "rocm6-diffusers" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-diffusers" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-diffusers + - localai/localai-backends:latest-gpu-amd-rocm-6-diffusers - !!merge <<: *diffusers name: "cuda11-diffusers" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-diffusers" @@ -1234,10 +1234,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-diffusers - !!merge <<: *diffusers - name: "rocm-diffusers-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-diffusers" + name: "rocm6-diffusers-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-diffusers" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-diffusers + - localai/localai-backends:master-gpu-amd-rocm-6-diffusers - !!merge <<: *diffusers name: "intel-diffusers-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-diffusers" @@ -1295,7 +1295,7 @@ capabilities: nvidia: "cuda12-kokoro-development" intel: "intel-kokoro-development" - amd: "rocm-kokoro-development" + amd-rocm-6: "rocm6-kokoro-development" nvidia-l4t: "nvidia-l4t-kokoro-development" - !!merge <<: *kokoro name: "cuda11-kokoro-development" @@ -1308,10 +1308,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-kokoro - !!merge <<: *kokoro - name: "rocm-kokoro-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-kokoro" + name: "rocm6-kokoro-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-kokoro" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-kokoro + - localai/localai-backends:master-gpu-amd-rocm-6-kokoro - !!merge <<: *kokoro name: "intel-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-kokoro" @@ -1343,10 +1343,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-kokoro - !!merge <<: *kokoro - name: "rocm-kokoro" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-kokoro" + name: "rocm6-kokoro" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-kokoro" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-kokoro + - localai/localai-backends:latest-gpu-amd-rocm-6-kokoro - !!merge <<: *kokoro name: "cuda13-kokoro" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-13-kokoro" @@ -1363,7 +1363,7 @@ capabilities: nvidia: "cuda12-faster-whisper-development" intel: "intel-faster-whisper-development" - amd: "rocm-faster-whisper-development" + amd-rocm-6: "rocm6-faster-whisper-development" nvidia-cuda-13: "cuda13-faster-whisper-development" - !!merge <<: *faster-whisper name: "cuda11-faster-whisper" @@ -1376,10 +1376,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-faster-whisper - !!merge <<: *faster-whisper - name: "rocm-faster-whisper-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-faster-whisper" + name: "rocm6-faster-whisper-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-faster-whisper" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-faster-whisper + - localai/localai-backends:master-gpu-amd-rocm-6-faster-whisper - !!merge <<: *faster-whisper name: "intel-faster-whisper" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-faster-whisper" @@ -1407,7 +1407,7 @@ capabilities: nvidia: "cuda12-coqui-development" intel: "intel-coqui-development" - amd: "rocm-coqui-development" + amd-rocm-6: "rocm6-coqui-development" - !!merge <<: *coqui name: "cuda11-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-nvidia-cuda-11-coqui" @@ -1429,10 +1429,10 @@ mirrors: - localai/localai-backends:master-gpu-nvidia-cuda-12-coqui - !!merge <<: *coqui - name: "rocm-coqui-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-coqui" + name: "rocm6-coqui-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-coqui" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-coqui + - localai/localai-backends:master-gpu-amd-rocm-6-coqui - !!merge <<: *coqui name: "intel-coqui" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-coqui" @@ -1444,17 +1444,17 @@ mirrors: - localai/localai-backends:master-gpu-intel-coqui - !!merge <<: *coqui - name: "rocm-coqui" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-coqui" + name: "rocm6-coqui" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-coqui" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-coqui + - localai/localai-backends:latest-gpu-amd-rocm-6-coqui ## bark - !!merge <<: *bark name: "bark-development" capabilities: nvidia: "cuda12-bark-development" intel: "intel-bark-development" - amd: "rocm-bark-development" + amd-rocm-6: "rocm6-bark-development" - !!merge <<: *bark name: "cuda11-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-11-bark" @@ -1466,10 +1466,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-11-bark - !!merge <<: *bark - name: "rocm-bark-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-bark" + name: "rocm6-bark-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-bark" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-bark + - localai/localai-backends:master-gpu-amd-rocm-6-bark - !!merge <<: *bark name: "intel-bark" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-bark" @@ -1486,10 +1486,10 @@ mirrors: - localai/localai-backends:latest-gpu-nvidia-cuda-12-bark - !!merge <<: *bark - name: "rocm-bark" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-bark" + name: "rocm6-bark" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-bark" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-bark + - localai/localai-backends:latest-gpu-amd-rocm-6-bark - !!merge <<: *bark name: "cuda12-bark-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-nvidia-cuda-12-bark" @@ -1587,7 +1587,7 @@ capabilities: nvidia: "cuda12-vibevoice-development" intel: "intel-vibevoice-development" - amd: "rocm-vibevoice-development" + amd-rocm-6: "rocm6-vibevoice-development" nvidia-l4t: "nvidia-l4t-vibevoice-development" default: "cpu-vibevoice-development" nvidia-cuda-13: "cuda13-vibevoice-development" @@ -1635,15 +1635,15 @@ mirrors: - localai/localai-backends:master-gpu-intel-vibevoice - !!merge <<: *vibevoice - name: "rocm-vibevoice" - uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-rocm-hipblas-vibevoice" + name: "rocm6-vibevoice" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-vibevoice" mirrors: - - localai/localai-backends:latest-gpu-rocm-hipblas-vibevoice + - localai/localai-backends:latest-gpu-amd-rocm-6-vibevoice - !!merge <<: *vibevoice - name: "rocm-vibevoice-development" - uri: "quay.io/go-skynet/local-ai-backends:master-gpu-rocm-hipblas-vibevoice" + name: "rocm6-vibevoice-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-vibevoice" mirrors: - - localai/localai-backends:master-gpu-rocm-hipblas-vibevoice + - localai/localai-backends:master-gpu-amd-rocm-6-vibevoice - !!merge <<: *vibevoice name: "nvidia-l4t-vibevoice" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-vibevoice" From cfeeff872938bea19e452cf33395cbcde84f66a5 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:07:39 -0500 Subject: [PATCH 18/20] Add backend/index.yaml entries for llama-cpp on rocm7 Signed-off-by: Simon Redman --- backend/index.yaml | 12 ++++++++++++ 1 file changed, 12 insertions(+) diff --git a/backend/index.yaml b/backend/index.yaml index 6042e9f8b6fc..2b6d367e624c 100644 --- a/backend/index.yaml +++ b/backend/index.yaml @@ -22,6 +22,7 @@ nvidia: "cuda12-llama-cpp" intel: "intel-sycl-f16-llama-cpp" amd-rocm-6: "rocm6-llama-cpp" + amd-rocm-7: "rocm7-llama-cpp" metal: "metal-llama-cpp" vulkan: "vulkan-llama-cpp" nvidia-l4t: "nvidia-l4t-arm64-llama-cpp" @@ -516,6 +517,7 @@ nvidia: "cuda12-llama-cpp-development" intel: "intel-sycl-f16-llama-cpp-development" amd-rocm-6: "rocm6-llama-cpp-development" + amd-rocm-7: "rocm7-llama-cpp-development" metal: "metal-llama-cpp-development" vulkan: "vulkan-llama-cpp-development" nvidia-l4t: "nvidia-l4t-arm64-llama-cpp-development" @@ -649,6 +651,11 @@ uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-llama-cpp" mirrors: - localai/localai-backends:latest-gpu-amd-rocm-6-llama-cpp +- !!merge <<: *llamacpp + name: "rocm7-llama-cpp" + uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-7-llama-cpp" + mirrors: + - localai/localai-backends:latest-gpu-amd-rocm-7-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp" uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-intel-sycl-f32-llama-cpp" @@ -694,6 +701,11 @@ uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-llama-cpp" mirrors: - localai/localai-backends:master-gpu-amd-rocm-6-llama-cpp +- !!merge <<: *llamacpp + name: "rocm7-llama-cpp-development" + uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-7-llama-cpp" + mirrors: + - localai/localai-backends:master-gpu-amd-rocm-7-llama-cpp - !!merge <<: *llamacpp name: "intel-sycl-f32-llama-cpp-development" uri: "quay.io/go-skynet/local-ai-backends:master-gpu-intel-sycl-f32-llama-cpp" From 3823f0d0632cb473581a812a1cc76f2c27ef82a7 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:20:29 -0500 Subject: [PATCH 19/20] Update docker tags for previously mis-named backends Signed-off-by: Simon Redman --- .github/workflows/backend.yml | 4 ++-- backend/index.yaml | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/.github/workflows/backend.yml b/.github/workflows/backend.yml index 6942c9b32cbb..6a0143a18bc0 100644 --- a/.github/workflows/backend.yml +++ b/.github/workflows/backend.yml @@ -1325,7 +1325,7 @@ jobs: platforms: 'linux/amd64' skip-drivers: 'true' tag-latest: 'auto' - tag-suffix: '-gpu-hipblas-exllama2' + tag-suffix: '-gpu-amd-rocm-6-exllama2' base-image: "rocm/dev-ubuntu-22.04:6.4.3" runs-on: 'ubuntu-latest' backend: "exllama2" @@ -1351,7 +1351,7 @@ jobs: # cuda-minor-version: "" # platforms: 'linux/amd64' # tag-latest: 'auto' - # tag-suffix: '-gpu-hipblas-rfdetr' + # tag-suffix: '-gpu-amd-rocm-6-rfdetr' # base-image: "rocm/dev-ubuntu-22.04:6.4.3" # runs-on: 'ubuntu-latest' # skip-drivers: 'false' diff --git a/backend/index.yaml b/backend/index.yaml index 2b6d367e624c..651f1bb20110 100644 --- a/backend/index.yaml +++ b/backend/index.yaml @@ -1005,9 +1005,9 @@ - localai/localai-backends:latest-gpu-intel-rfdetr # - !!merge <<: *rfdetr # name: "rocm6-rfdetr" -# uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-hipblas-rfdetr" +# uri: "quay.io/go-skynet/local-ai-backends:latest-gpu-amd-rocm-6-rfdetr" # mirrors: -# - localai/localai-backends:latest-gpu-hipblas-rfdetr +# - localai/localai-backends:latest-gpu-amd-rocm-6-rfdetr - !!merge <<: *rfdetr name: "nvidia-l4t-arm64-rfdetr" uri: "quay.io/go-skynet/local-ai-backends:latest-nvidia-l4t-arm64-rfdetr" @@ -1035,9 +1035,9 @@ - localai/localai-backends:master-gpu-intel-rfdetr # - !!merge <<: *rfdetr # name: "rocm6-rfdetr-development" -# uri: "quay.io/go-skynet/local-ai-backends:master-gpu-hipblas-rfdetr" +# uri: "quay.io/go-skynet/local-ai-backends:master-gpu-amd-rocm-6-rfdetr" # mirrors: -# - localai/localai-backends:master-gpu-hipblas-rfdetr +# - localai/localai-backends:master-gpu-amd-rocm-6-rfdetr - !!merge <<: *rfdetr name: "cpu-rfdetr-development" uri: "quay.io/go-skynet/local-ai-backends:master-cpu-rfdetr" From 4941f1396f696891e04ea6e28569f5f7e76b0a16 Mon Sep 17 00:00:00 2001 From: Simon Redman Date: Sun, 21 Dec 2025 16:21:12 -0500 Subject: [PATCH 20/20] Bulk update documentation with new image tag names Signed-off-by: Simon Redman --- README.md | 14 +++++++------- docs/content/features/GPU-acceleration.md | 4 ++-- docs/content/getting-started/container-images.md | 10 +++++----- docs/content/installation/docker.md | 6 +++--- docs/static/install.sh | 4 ++-- 5 files changed, 19 insertions(+), 19 deletions(-) diff --git a/README.md b/README.md index d67886a116ac..7e031ff435ee 100644 --- a/README.md +++ b/README.md @@ -43,7 +43,7 @@ > :bulb: Get help - [❓FAQ](https://localai.io/faq/) [πŸ’­Discussions](https://github.com/go-skynet/LocalAI/discussions) [:speech_balloon: Discord](https://discord.gg/uJAeKSAGDy) [:book: Documentation website](https://localai.io/) > -> [πŸ’» Quickstart](https://localai.io/basics/getting_started/) [πŸ–ΌοΈ Models](https://models.localai.io/) [πŸš€ Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [πŸ›« Examples](https://github.com/mudler/LocalAI-examples) Try on +> [πŸ’» Quickstart](https://localai.io/basics/getting_started/) [πŸ–ΌοΈ Models](https://models.localai.io/) [πŸš€ Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [πŸ›« Examples](https://github.com/mudler/LocalAI-examples) Try on [![Telegram](https://img.shields.io/badge/Telegram-2CA5E0?style=for-the-badge&logo=telegram&logoColor=white)](https://t.me/localaiofficial_bot) [![tests](https://github.com/go-skynet/LocalAI/actions/workflows/test.yml/badge.svg)](https://github.com/go-skynet/LocalAI/actions/workflows/test.yml)[![Build and Release](https://github.com/go-skynet/LocalAI/actions/workflows/release.yaml/badge.svg)](https://github.com/go-skynet/LocalAI/actions/workflows/release.yaml)[![build container images](https://github.com/go-skynet/LocalAI/actions/workflows/image.yml/badge.svg)](https://github.com/go-skynet/LocalAI/actions/workflows/image.yml)[![Bump dependencies](https://github.com/go-skynet/LocalAI/actions/workflows/bump_deps.yaml/badge.svg)](https://github.com/go-skynet/LocalAI/actions/workflows/bump_deps.yaml)[![Artifact Hub](https://img.shields.io/endpoint?url=https://artifacthub.io/badge/repository/localai)](https://artifacthub.io/packages/search?repo=localai) @@ -131,10 +131,10 @@ For more installation options, see [Installer Options](https://localai.io/instal Or run with docker: > **πŸ’‘ Docker Run vs Docker Start** -> +> > - `docker run` creates and starts a new container. If a container with the same name already exists, this command will fail. > - `docker start` starts an existing container that was previously created with `docker run`. -> +> > If you've already run LocalAI before and want to start it again, use: `docker start -i local-ai` ### CPU only image: @@ -159,7 +159,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-nv ### AMD GPU Images (ROCm): ```bash -docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-hipblas +docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-amd-rocm-6 ``` ### Intel GPU Images (oneAPI): @@ -190,7 +190,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-ai docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel # AMD GPU version -docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-hipblas +docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-amd-rocm-6 ``` For more information about the AIO images and pre-downloaded models, see [Container Documentation](https://localai.io/basics/container/). @@ -250,7 +250,7 @@ Roadmap items: [List of issues](https://github.com/mudler/LocalAI/issues?q=is%3A - πŸ—£ [Text to Audio](https://localai.io/features/text-to-audio/) - πŸ”ˆ [Audio to Text](https://localai.io/features/audio-to-text/) (Audio transcription with `whisper.cpp`) - 🎨 [Image generation](https://localai.io/features/image-generation) -- πŸ”₯ [OpenAI-alike tools API](https://localai.io/features/openai-functions/) +- πŸ”₯ [OpenAI-alike tools API](https://localai.io/features/openai-functions/) - 🧠 [Embeddings generation for vector databases](https://localai.io/features/embeddings/) - ✍️ [Constrained grammars](https://localai.io/features/constrained_grammars/) - πŸ–ΌοΈ [Download Models directly from Huggingface ](https://localai.io/models/) @@ -356,7 +356,7 @@ Other: - Github bot which answer on issues, with code and documentation as context https://github.com/JackBekket/GitHelper - Github Actions: https://github.com/marketplace/actions/start-localai - Examples: https://github.com/mudler/LocalAI/tree/master/examples/ - + ### πŸ”— Resources diff --git a/docs/content/features/GPU-acceleration.md b/docs/content/features/GPU-acceleration.md index 7c619962b450..7e1d0e197b62 100644 --- a/docs/content/features/GPU-acceleration.md +++ b/docs/content/features/GPU-acceleration.md @@ -181,7 +181,7 @@ The following are examples of the ROCm specific configuration elements required. ```yaml # For full functionality select a non-'core' image, version locking the image is recommended for debug purposes. - image: quay.io/go-skynet/local-ai:master-aio-gpu-hipblas + image: quay.io/go-skynet/local-ai:master-aio-gpu-amd-rocm-6 environment: - DEBUG=true # If your gpu is not already included in the current list of default targets the following build details are required. @@ -204,7 +204,7 @@ docker run \ -e GPU_TARGETS=gfx906 \ --device /dev/dri \ --device /dev/kfd \ - quay.io/go-skynet/local-ai:master-aio-gpu-hipblas + quay.io/go-skynet/local-ai:master-aio-gpu-amd-rocm-6 ``` Please ensure to add all other required environment variables, port forwardings, etc to your `compose` file or `run` command. diff --git a/docs/content/getting-started/container-images.md b/docs/content/getting-started/container-images.md index a4e94d8b8f05..967157f3f1e2 100644 --- a/docs/content/getting-started/container-images.md +++ b/docs/content/getting-started/container-images.md @@ -84,9 +84,9 @@ Standard container images do not have pre-installed models. Use these if you wan | Description | Quay | Docker Hub | | --- | --- |-------------------------------------------------------------| -| Latest images from the branch (development) | `quay.io/go-skynet/local-ai:master-gpu-hipblas` | `localai/localai:master-gpu-hipblas` | -| Latest tag | `quay.io/go-skynet/local-ai:latest-gpu-hipblas` | `localai/localai:latest-gpu-hipblas` | -| Versioned image | `quay.io/go-skynet/local-ai:{{< version >}}-gpu-hipblas` | `localai/localai:{{< version >}}-gpu-hipblas` | +| Latest images from the branch (development) | `quay.io/go-skynet/local-ai:master-gpu-amd-rocm-6` | `localai/localai:master-gpu-amd-rocm-6` | +| Latest tag | `quay.io/go-skynet/local-ai:latest-gpu-amd-rocm-6` | `localai/localai:latest-gpu-amd-rocm-6` | +| Versioned image | `quay.io/go-skynet/local-ai:{{< version >}}-gpu-amd-rocm-6` | `localai/localai:{{< version >}}-gpu-amd-rocm-6` | {{% /tab %}} @@ -178,7 +178,7 @@ services: **Models caching**: The **AIO** image will download the needed models on the first run if not already present and store those in `/models` inside the container. The AIO models will be automatically updated with new versions of AIO images. -You can change the directory inside the container by specifying a `MODELS_PATH` environment variable (or `--models-path`). +You can change the directory inside the container by specifying a `MODELS_PATH` environment variable (or `--models-path`). If you want to use a named model or a local directory, you can mount it as a volume to `/models`: @@ -203,7 +203,7 @@ docker run -p 8080:8080 --name local-ai -ti -v localai-models:/models localai/lo | Versioned image (e.g. for CPU) | `quay.io/go-skynet/local-ai:{{< version >}}-aio-cpu` | `localai/localai:{{< version >}}-aio-cpu` | | Latest images for Nvidia GPU (CUDA11) | `quay.io/go-skynet/local-ai:latest-aio-gpu-nvidia-cuda-11` | `localai/localai:latest-aio-gpu-nvidia-cuda-11` | | Latest images for Nvidia GPU (CUDA12) | `quay.io/go-skynet/local-ai:latest-aio-gpu-nvidia-cuda-12` | `localai/localai:latest-aio-gpu-nvidia-cuda-12` | -| Latest images for AMD GPU | `quay.io/go-skynet/local-ai:latest-aio-gpu-hipblas` | `localai/localai:latest-aio-gpu-hipblas` | +| Latest images for AMD GPU | `quay.io/go-skynet/local-ai:latest-aio-gpu-amd-rocm-6` | `localai/localai:latest-aio-gpu-amd-rocm-6` | | Latest images for Intel GPU | `quay.io/go-skynet/local-ai:latest-aio-gpu-intel` | `localai/localai:latest-aio-gpu-intel` | ### Available environment variables diff --git a/docs/content/installation/docker.md b/docs/content/installation/docker.md index 125968d28ce2..85b662c96ad4 100644 --- a/docs/content/installation/docker.md +++ b/docs/content/installation/docker.md @@ -70,7 +70,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gp **AMD GPU (ROCm):** ```bash -docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-hipblas +docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-gpu-amd-rocm-6 ``` **Intel GPU:** @@ -112,7 +112,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-ai **AMD GPU (ROCm):** ```bash -docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-hipblas +docker run -ti --name local-ai -p 8080:8080 --device=/dev/kfd --device=/dev/dri --group-add=video localai/localai:latest-aio-gpu-amd-rocm-6 ``` **Intel GPU:** @@ -132,7 +132,7 @@ services: # For GPU support, use one of: # image: localai/localai:latest-aio-gpu-nvidia-cuda-12 # image: localai/localai:latest-aio-gpu-nvidia-cuda-11 - # image: localai/localai:latest-aio-gpu-hipblas + # image: localai/localai:latest-aio-gpu-amd-rocm-6 # image: localai/localai:latest-aio-gpu-intel healthcheck: test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] diff --git a/docs/static/install.sh b/docs/static/install.sh index 8dfe7e4a1111..00fbb8253592 100755 --- a/docs/static/install.sh +++ b/docs/static/install.sh @@ -702,10 +702,10 @@ install_docker() { $envs \ -d -p $PORT:8080 --name local-ai localai/localai:$IMAGE_TAG $STARTCOMMAND elif [ "$HAS_AMD" ]; then - IMAGE_TAG=${LOCALAI_VERSION}-gpu-hipblas + IMAGE_TAG=${LOCALAI_VERSION}-gpu-amd-rocm-6 # AIO if [ "$USE_AIO" = true ]; then - IMAGE_TAG=${LOCALAI_VERSION}-aio-gpu-hipblas + IMAGE_TAG=${LOCALAI_VERSION}-aio-gpu-amd-rocm-6 fi info "Starting LocalAI Docker container..."