public

riva-speech (2.19.0)

Published 2025-07-02 05:09:06 +00:00 by publish

Installation

docker pull packages.speedcloud.co.in/public/riva-speech:2.19.0
sha256:2060fea00fda41d11dc82dc084814ee5abe196669efd47da299b9a02ba9ff00c

Image Layers

ARG RELEASE
ARG LAUNCHPAD_BUILD_ARCH
LABEL org.opencontainers.image.ref.name=ubuntu
LABEL org.opencontainers.image.version=24.04
ADD file:bcebbf0fddcba5b864d5d267b68dd23bcfb01275e6ec7bcab69bf8b56af14804 in /
CMD ["/bin/bash"]
RUN /bin/sh -c export DEBIAN_FRONTEND=noninteractive && apt-get update && apt-get install -y --no-install-recommends apt-utils build-essential ca-certificates curl libncurses6 libncursesw6 patch wget rsync unzip jq gnupg libtcmalloc-minimal4 && rm -rf /var/lib/apt/lists/* && echo "hsts=0" > /root/.wgetrc # buildkit
ARG CUDA_VERSION=12.8.0.038
ARG CUDA_DRIVER_VERSION=570.86.10
ARG JETPACK_HOST_MOUNTS=
ENV CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 CUDA_CACHE_DISABLE=1 NVIDIA_REQUIRE_JETPACK_HOST_MOUNTS=
RUN |3 CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 JETPACK_HOST_MOUNTS= /bin/sh -c if [ -n "${JETPACK_HOST_MOUNTS}" ]; then echo "/usr/lib/aarch64-linux-gnu/tegra" > /etc/ld.so.conf.d/nvidia-tegra.conf && echo "/usr/lib/aarch64-linux-gnu/tegra-egl" >> /etc/ld.so.conf.d/nvidia-tegra.conf; fi # buildkit
RUN |3 CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 JETPACK_HOST_MOUNTS= /bin/sh -c /nvidia/build-scripts/installCUDA.sh # buildkit
RUN |3 CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 JETPACK_HOST_MOUNTS= /bin/sh -c cp -vprd /nvidia/. / && patch -p0 < /etc/startup_scripts.patch && rm -f /etc/startup_scripts.patch # buildkit
ENV _CUDA_COMPAT_PATH=/usr/local/cuda/compat ENV=/etc/shinit_v2 BASH_ENV=/etc/bash.bashrc SHELL=/bin/bash NVIDIA_REQUIRE_CUDA=cuda>=9.0
LABEL com.nvidia.volumes.needed=nvidia_driver com.nvidia.cuda.version=9.0
ARG NCCL_VERSION=2.25.1
ARG CUBLAS_VERSION=12.8.3.14
ARG CUFFT_VERSION=11.3.3.41
ARG CURAND_VERSION=10.3.9.55
ARG CUSPARSE_VERSION=12.5.7.53
ARG CUSOLVER_VERSION=11.7.2.55
ARG CUTENSOR_VERSION=2.1.0.9
ARG NPP_VERSION=12.3.3.65
ARG NVJPEG_VERSION=12.3.5.57
ARG CUFILE_VERSION=1.13.0.11
ARG NVJITLINK_VERSION=12.8.61
ARG CUDNN_VERSION=9.7.0.66
ARG CUDNN_FRONTEND_VERSION=1.9.0
ARG TRT_VERSION=10.8.0.43
ARG TRTOSS_VERSION=
ARG NSIGHT_SYSTEMS_VERSION=2024.6.2.225
ARG NSIGHT_COMPUTE_VERSION=2025.1.0.14
ARG CUSPARSELT_VERSION=0.6.3.2
ENV NCCL_VERSION=2.25.1 CUBLAS_VERSION=12.8.3.14 CUFFT_VERSION=11.3.3.41 CURAND_VERSION=10.3.9.55 CUSPARSE_VERSION=12.5.7.53 CUSPARSELT_VERSION=0.6.3.2 CUSOLVER_VERSION=11.7.2.55 CUTENSOR_VERSION=2.1.0.9 NPP_VERSION=12.3.3.65 NVJPEG_VERSION=12.3.5.57 CUFILE_VERSION=1.13.0.11 NVJITLINK_VERSION=12.8.61 CUDNN_VERSION=9.7.0.66 CUDNN_FRONTEND_VERSION=1.9.0 TRT_VERSION=10.8.0.43 TRTOSS_VERSION= NSIGHT_SYSTEMS_VERSION=2024.6.2.225 NSIGHT_COMPUTE_VERSION=2025.1.0.14
RUN |21 CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 JETPACK_HOST_MOUNTS= NCCL_VERSION=2.25.1 CUBLAS_VERSION=12.8.3.14 CUFFT_VERSION=11.3.3.41 CURAND_VERSION=10.3.9.55 CUSPARSE_VERSION=12.5.7.53 CUSOLVER_VERSION=11.7.2.55 CUTENSOR_VERSION=2.1.0.9 NPP_VERSION=12.3.3.65 NVJPEG_VERSION=12.3.5.57 CUFILE_VERSION=1.13.0.11 NVJITLINK_VERSION=12.8.61 CUDNN_VERSION=9.7.0.66 CUDNN_FRONTEND_VERSION=1.9.0 TRT_VERSION=10.8.0.43 TRTOSS_VERSION= NSIGHT_SYSTEMS_VERSION=2024.6.2.225 NSIGHT_COMPUTE_VERSION=2025.1.0.14 CUSPARSELT_VERSION=0.6.3.2 /bin/sh -c /nvidia/build-scripts/installLIBS.sh && /nvidia/build-scripts/installCUDNN.sh && /nvidia/build-scripts/installTRT.sh && /nvidia/build-scripts/installNSYS.sh && /nvidia/build-scripts/installNCU.sh && /nvidia/build-scripts/installCUTENSOR.sh && /nvidia/build-scripts/installCUSPARSELT.sh && if [ -z "${JETPACK_HOST_MOUNTS}" ]; then /nvidia/build-scripts/installNCCL.sh; fi; # buildkit
LABEL com.nvidia.nccl.version=2.25.1 com.nvidia.cublas.version=12.8.3.14 com.nvidia.cufft.version=11.3.3.41 com.nvidia.curand.version=10.3.9.55 com.nvidia.cusparse.version=12.5.7.53 com.nvidia.cusparselt.version=0.6.3.2 com.nvidia.cusolver.version=11.7.2.55 com.nvidia.cutensor.version=2.1.0.9 com.nvidia.npp.version=12.3.3.65 com.nvidia.nvjpeg.version=12.3.5.57 com.nvidia.cudnn.version=9.7.0.66 com.nvidia.tensorrt.version=10.8.0.43 com.nvidia.tensorrtoss.version= com.nvidia.nsightsystems.version=2024.6.2.225 com.nvidia.nsightcompute.version=2025.1.0.14
ARG DALI_VERSION=1.45.0
ARG DALI_BUILD=
ARG DALI_URL_SUFFIX=120
ARG POLYGRAPHY_VERSION=0.49.16
ARG TRANSFORMER_ENGINE_VERSION=1.14
ARG MODEL_OPT_VERSION=0.21.0
ENV DALI_VERSION=1.45.0 DALI_BUILD= DALI_URL_SUFFIX=120 POLYGRAPHY_VERSION=0.49.16 TRANSFORMER_ENGINE_VERSION=1.14 MODEL_OPT_VERSION=0.21.0
ADD docs.tgz / # buildkit
RUN |27 CUDA_VERSION=12.8.0.038 CUDA_DRIVER_VERSION=570.86.10 JETPACK_HOST_MOUNTS= NCCL_VERSION=2.25.1 CUBLAS_VERSION=12.8.3.14 CUFFT_VERSION=11.3.3.41 CURAND_VERSION=10.3.9.55 CUSPARSE_VERSION=12.5.7.53 CUSOLVER_VERSION=11.7.2.55 CUTENSOR_VERSION=2.1.0.9 NPP_VERSION=12.3.3.65 NVJPEG_VERSION=12.3.5.57 CUFILE_VERSION=1.13.0.11 NVJITLINK_VERSION=12.8.61 CUDNN_VERSION=9.7.0.66 CUDNN_FRONTEND_VERSION=1.9.0 TRT_VERSION=10.8.0.43 TRTOSS_VERSION= NSIGHT_SYSTEMS_VERSION=2024.6.2.225 NSIGHT_COMPUTE_VERSION=2025.1.0.14 CUSPARSELT_VERSION=0.6.3.2 DALI_VERSION=1.45.0 DALI_BUILD= DALI_URL_SUFFIX=120 POLYGRAPHY_VERSION=0.49.16 TRANSFORMER_ENGINE_VERSION=1.14 MODEL_OPT_VERSION=0.21.0 /bin/sh -c echo "/usr/local/nvidia/lib" >> /etc/ld.so.conf.d/nvidia.conf && echo "/usr/local/nvidia/lib64" >> /etc/ld.so.conf.d/nvidia.conf # buildkit
ARG _LIBPATH_SUFFIX=
ENV PATH=/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin LD_LIBRARY_PATH=/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64 NVIDIA_VISIBLE_DEVICES=all NVIDIA_DRIVER_CAPABILITIES=compute,utility,video
COPY entrypoint/ /opt/nvidia/ # buildkit
ENV NVIDIA_PRODUCT_NAME=CUDA
ENTRYPOINT ["/opt/nvidia/nvidia_entrypoint.sh"]
COPY NVIDIA_Deep_Learning_Container_License.pdf /workspace/ # buildkit
RUN /bin/sh -c export DEBIAN_FRONTEND=noninteractive && apt-get update && apt-get install -y --no-install-recommends build-essential git libglib2.0-0 less libhwloc15 libnl-route-3-200 libnl-3-dev libnl-route-3-dev libnuma-dev libnuma1 libpmi2-0-dev nano numactl openssh-client vim wget && rm -rf /var/lib/apt/lists/* # buildkit
ARG GDRCOPY_VERSION=2.4.1
ARG HPCX_VERSION=2.21
ARG RDMACORE_VERSION=39.0
ARG MOFED_VERSION=5.4-rdmacore39.0
ARG OPENUCX_VERSION=1.18.0
ARG OPENMPI_VERSION=4.1.7
ARG EFA_VERSION=1.34.0
ARG AWS_OFI_NCCL_VERSION=1.12.1
ENV GDRCOPY_VERSION=2.4.1 HPCX_VERSION=2.21 MOFED_VERSION=5.4-rdmacore39.0 OPENUCX_VERSION=1.18.0 OPENMPI_VERSION=4.1.7 RDMACORE_VERSION=39.0 EFA_VERSION=1.34.0 AWS_OFI_NCCL_VERSION=1.12.1
ARG TARGETARCH=amd64
RUN |9 GDRCOPY_VERSION=2.4.1 HPCX_VERSION=2.21 RDMACORE_VERSION=39.0 MOFED_VERSION=5.4-rdmacore39.0 OPENUCX_VERSION=1.18.0 OPENMPI_VERSION=4.1.7 EFA_VERSION=1.34.0 AWS_OFI_NCCL_VERSION=1.12.1 TARGETARCH=amd64 /bin/sh -c cd /nvidia && ( export DEBIAN_FRONTEND=noninteractive && apt-get update && apt-get install -y --no-install-recommends libibverbs1 libibverbs-dev librdmacm1 librdmacm-dev libibumad3 libibumad-dev ibverbs-utils ibverbs-providers && rm -rf /var/lib/apt/lists/* && rm $(dpkg-query -L libibverbs-dev librdmacm-dev libibumad-dev | grep "\(\.so\|\.a\)$") ) && ( cd opt/gdrcopy/ && dpkg -i libgdrapi_*.deb ) && ( cp -r opt/hpcx /opt/ && cp etc/ld.so.conf.d/hpcx.conf /etc/ld.so.conf.d/ && ln -sf /opt/hpcx/ompi /usr/local/mpi && ln -sf /opt/hpcx/ucx /usr/local/ucx && sed -i 's/^\(hwloc_base_binding_policy\) = core$/\1 = none/' /opt/hpcx/ompi/etc/openmpi-mca-params.conf && sed -i 's/^\(btl = self\)$/#\1/' /opt/hpcx/ompi/etc/openmpi-mca-params.conf ) && ( if [ ! -f /etc/ld.so.conf.d/nvidia-tegra.conf ]; then cd opt/amazon/efa/ && dpkg -i libfabric*.deb && rm /opt/amazon/efa/lib/libfabric.a && echo "/opt/amazon/efa/lib" > /etc/ld.so.conf.d/efa.conf; fi ) && ldconfig # buildkit
ENV OPAL_PREFIX=/opt/hpcx/ompi PATH=/usr/local/mpi/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/ucx/bin:/opt/amazon/efa/bin
ENV OMPI_MCA_coll_hcoll_enable=0
ENV NCCL_NVLS_ENABLE=0
COPY cuda-*.patch /tmp # buildkit
RUN |9 GDRCOPY_VERSION=2.4.1 HPCX_VERSION=2.21 RDMACORE_VERSION=39.0 MOFED_VERSION=5.4-rdmacore39.0 OPENUCX_VERSION=1.18.0 OPENMPI_VERSION=4.1.7 EFA_VERSION=1.34.0 AWS_OFI_NCCL_VERSION=1.12.1 TARGETARCH=amd64 /bin/sh -c export DEVEL=1 BASE=0 && /nvidia/build-scripts/installNCU.sh && /nvidia/build-scripts/installCUDA.sh && /nvidia/build-scripts/installLIBS.sh && if [ ! -f /etc/ld.so.conf.d/nvidia-tegra.conf ]; then /nvidia/build-scripts/installNCCL.sh; fi && /nvidia/build-scripts/installCUDNN.sh && /nvidia/build-scripts/installCUTENSOR.sh && /nvidia/build-scripts/installTRT.sh && /nvidia/build-scripts/installNSYS.sh && /nvidia/build-scripts/installCUSPARSELT.sh && if [ -f "/tmp/cuda-${_CUDA_VERSION_MAJMIN}.patch" ]; then patch -p0 < /tmp/cuda-${_CUDA_VERSION_MAJMIN}.patch; fi && rm -f /tmp/cuda-*.patch # buildkit
ENV LIBRARY_PATH=/usr/local/cuda/lib64/stubs:
COPY /opt/amazon/aws-ofi-nccl /opt/amazon/aws-ofi-nccl # buildkit
RUN |9 GDRCOPY_VERSION=2.4.1 HPCX_VERSION=2.21 RDMACORE_VERSION=39.0 MOFED_VERSION=5.4-rdmacore39.0 OPENUCX_VERSION=1.18.0 OPENMPI_VERSION=4.1.7 EFA_VERSION=1.34.0 AWS_OFI_NCCL_VERSION=1.12.1 TARGETARCH=amd64 /bin/sh -c if [ ! -f /etc/ld.so.conf.d/nvidia-tegra.conf ]; then echo "/opt/amazon/aws-ofi-nccl/lib" > /etc/ld.so.conf.d/aws-ofi-nccl.conf && ldconfig; fi # buildkit
ARG NVIDIA_TRITON_SERVER_BASE_VERSION
ARG NVIDIA_BUILD_ID
ENV NVIDIA_TRITON_SERVER_BASE_VERSION=25.01
ENV NVIDIA_BUILD_ID=136230209
ENV NVIDIA_PRODUCT_NAME=Triton Server Base
COPY docker/entrypoint.d/ /opt/nvidia/entrypoint.d/ # buildkit
RUN |2 NVIDIA_TRITON_SERVER_BASE_VERSION=25.01 NVIDIA_BUILD_ID=136230209 /bin/sh -c sed 's/Server/Server Base/' /opt/nvidia/entrypoint.d/10-banner.txt | sed 's/^===/========/' > /opt/nvidia/entrypoint.d/10-banner.new && mv /opt/nvidia/entrypoint.d/10-banner.new /opt/nvidia/entrypoint.d/10-banner.txt # buildkit
ARG DEBIAN_FRONTEND=noninteractive
ARG TARGET_ARCH=x86_64 # valid values: x86_64, aarch64
ARG TARGET_OS=linux # valid values: linux, l4t, qnx
ARG BUILD_MODE=release # valid values: release, debug
ARG TRITON_RELEASE_VERSION=2.54.0
ARG TRITON_RELEASE_VERSION_L4T=2.50.0
ARG DCGM_VERSION=3.2.6
ARG TARGET_CUDA_VERSION=12.8
WORKDIR /opt/tritonserver
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c apt-get update && apt-get install -y --no-install-recommends software-properties-common python3-dev python3-pip wget && rm -rf /var/lib/apt/lists/* # buildkit
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_OS" != "l4t" ] && [ "$TARGET_OS" != "qnx" ]; then cp -R /opt2/tritonserver/bin . && cp -R /opt2/tritonserver/lib . && cp -R /opt2/tritonserver/include . && cp /opt2/tritonserver/TRITON_VERSION . && mkdir /opt/tritonserver/backends && cp -R /opt2/tritonserver/backends/tensorrt backends/tensorrt && cp -R /opt2/tritonserver/backends/onnxruntime backends/onnxruntime && cp -R /opt2/tritonserver/backends/python backends/python && cp -R /opt2/tritonserver/backends/pytorch backends/pytorch; else cd /opt && wget -nv https://github.com/triton-inference-server/server/releases/download/v${TRITON_RELEASE_VERSION_L4T}/tritonserver-${TRITON_RELEASE_VERSION_L4T}-igpu.tar.gz && tar -zxf tritonserver-${TRITON_RELEASE_VERSION_L4T}-igpu.tar.gz tritonserver/bin tritonserver/lib tritonserver/include tritonserver/TRITON_VERSION tritonserver/backends/onnxruntime tritonserver/backends/tensorrt tritonserver/backends/python tritonserver/backends/pytorch && rm tritonserver-${TRITON_RELEASE_VERSION_L4T}-igpu.tar.gz; fi # buildkit
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_OS" != "l4t" ] && [ "$TARGET_OS" != "qnx" ]; then cp -R /pytorch_lib/python3.12/dist-packages/torch/include /opt/tritonserver/backends/pytorch/include; else cp -R /pytorch_lib/python3.10/dist-packages/torch/include /opt/tritonserver/backends/pytorch/include; fi # buildkit
SHELL [/bin/bash -c]
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/bash -c if [ "$TARGET_OS" != "l4t" ] && [ "$TARGET_OS" != "qnx" ]; then wget -nv https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/${TARGET_ARCH//aarch64/sbsa}/cuda-ubuntu2204.pin && mv cuda-ubuntu2204.pin /etc/apt/preferences.d/cuda-repository-pin-600 && wget -nv https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/${TARGET_ARCH//aarch64/sbsa}/cuda-keyring_1.0-1_all.deb && dpkg -i cuda-keyring_1.0-1_all.deb && add-apt-repository "deb https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/${TARGET_ARCH//aarch64/sbsa}/ /" && apt-get update && apt-get install -y datacenter-gpu-manager=1:${DCGM_VERSION} && rm -f cuda-keyring_1.0-1_all.deb; fi && apt-get update && apt-get install -y --no-install-recommends libboost-program-options1.74.0 libboost-system1.74.0 libboost-thread1.74.0 libboost-locale1.74.0 && wget -nv -O /tmp/boost.tar.gz https://archives.boost.io/release/1.79.0/source/boost_1_79_0.tar.gz && tar xzf /tmp/boost.tar.gz -C /tmp/ && cd /tmp/boost_1_79_0 && ./bootstrap.sh --prefix=/usr && ./b2 install && rm -rf /tmp/boost* && wget -nv -O /tmp/boost.tar.gz https://archives.boost.io/release/1.80.0/source/boost_1_80_0.tar.gz && tar xzf /tmp/boost.tar.gz -C /tmp/ && cd /tmp/boost_1_80_0 && ./bootstrap.sh --prefix=/usr && ./b2 install && mv /tmp/boost_1_80_0/boost /usr/include/boost && rm -rf /tmp/boost* && cd /tmp && wget -nv https://urm.nvidia.com/artifactory/sw-ai-app-bazel/sw-ai-app-jarvis/tritonserver-be/onnxruntime-24.06/onnxruntime-backend-${TARGET_ARCH//x86_64/amd64}.tar && rm -rf /opt/tritonserver/backends/onnxruntime && tar -xvf /tmp/onnxruntime-backend-${TARGET_ARCH//x86_64/amd64}.tar -C /opt/tritonserver/backends && rm -rf /tmp/onnxruntime* && apt-get install -y libcudnn8-dev # buildkit
SHELL [/bin/sh -c]
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c python3 -m pip config set global.break-system-packages true # buildkit
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_ARCH" = "x86_64" ]; then wget -nv https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/tars/TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-12.8.tar.gz && tar xf TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-12.8.tar.gz && pip3 install --no-cache-dir --force-reinstall TensorRT-10.8.0.43/python/tensorrt-10.8.0.43-cp312-none-linux_x86_64.whl && rm -rf TensorRT*; elif [ "$TARGET_ARCH" = "aarch64" ] && [ "$TARGET_OS" = "linux" ]; then wget -nv https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/tars/TensorRT-10.8.0.43.Linux.aarch64-gnu.cuda-12.8.tar.gz && tar xf TensorRT-10.8.0.43.Linux.aarch64-gnu.cuda-12.8.tar.gz && pip3 install --no-cache-dir --force-reinstall TensorRT-10.8.0.43/python/tensorrt-10.8.0.43-cp312-none-linux_aarch64.whl && rm -rf TensorRT*; elif [ "$TARGET_ARCH" = "aarch64" ] && [ "$TARGET_OS" = "l4t" ]; then wget -nv https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/tars/TensorRT-10.6.0.26.l4t.aarch64-gnu.cuda-12.6.tar.gz && tar xf TensorRT-10.6.0.26.l4t.aarch64-gnu.cuda-12.6.tar.gz && pip3 install --no-cache-dir --force-reinstall TensorRT-10.6.0.26/python/tensorrt-10.6.0-cp310-none-linux_aarch64.whl && wget -nv https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/arm64/cuda-keyring_1.1-1_all.deb && dpkg -i cuda-keyring_1.1-1_all.deb && apt-get update && apt-get install -y libcudnn9-dev libnvinfer-dev && rm -rf TensorRT* cuda-keyring*; elif [ "$TARGET_ARCH" = "aarch64" ] && [ "$TARGET_OS" = "qnx" ]; then apt-get purge -y libcudnn8* libnvinfer* && wget -nv http://cuda-repo/release-candidates/kitpicks/cudnn-v8-4-cuda-11-4-orin/8.4.1.88/001/redist/cudnn/cudnn/linux-aarch64/cudnn-linux-aarch64-8.4.1.88_cuda11.4-archive.tar.xz && tar xvf cudnn-linux-aarch64-8.4.1.88_cuda11.4-archive.tar.xz && mv cudnn-linux-aarch64-8.4.1.88_cuda11.4-archive cudnn && rm -rf /lib/aarch64-linux-gnu/libcudnn* /usr/include/aarch64-linux-gnu/cudnn* && cp -r cudnn/lib/* /lib/aarch64-linux-gnu/ && cp -r cudnn/include/* /usr/include/aarch64-linux-gnu/ && rm -rf cudnn cudnn-linux-aarch64-8.4.1.88_cuda11.4-archive.tar.xz && wget -nv http://cuda-repo.nvidia.com/release-candidates/Libraries/TensorRT/v8.4/8.4.11.6-3306d99a/11.4-r470/d6l-aarch64/tar/TensorRT-8.4.11.6.Ubuntu-20.04.aarch64-gnu.cuda-11.4.cudnn8.4.tar.gz && tar xvf TensorRT-8.4.11.6.Ubuntu-20.04.aarch64-gnu.cuda-11.4.cudnn8.4.tar.gz && mv TensorRT-8.4.11.6 TensorRT && cp -r TensorRT/lib/* /lib/aarch64-linux-gnu/ && cp -r TensorRT/include/* /usr/include/aarch64-linux-gnu/ && rm -rf TensorRT TensorRT-8.4.11.6.Ubuntu-20.04.aarch64-gnu.cuda-11.4.cudnn8.4.tar.gz; fi # buildkit
RUN |24 DEBIAN_FRONTEND=noninteractive TARGET_ARCH=x86_64 #= valid= values:= x86_64,= aarch64= TARGET_OS=linux #= valid= values:= linux,= l4t,= qnx= BUILD_MODE=release #= valid= values:= release,= debug= TRITON_RELEASE_VERSION=2.54.0 TRITON_RELEASE_VERSION_L4T=2.50.0 DCGM_VERSION=3.2.6 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_ARCH" = "x86_64" ]; then pip3 install --no-cache-dir https://urm.nvidia.com/artifactory/sw-ai-app-bazel-local/sw-ai-app-jarvis/tools/torch-2.6.0a0+ecf3bae40a.nv25.1-22181419-cp312-cp312-linux_x86_64.whl && pip3 install --no-cache-dir https://gitlab-master.nvidia.com/api/v4/projects/100660/packages/pypi/files/5fab58d043583f9b59df5c36bb797fd147cc488fff1c79fc11e974721845ea9e/torchvision-0.20.0a0+ecf3bae40a.nv25.1-22932737-cp312-cp312-manylinux_2_28_x86_64.whl#sha256=5fab58d043583f9b59df5c36bb797fd147cc488fff1c79fc11e974721845ea9e && pip3 install --no-cache-dir tensorrt_cu12==10.8.0.43 --pre --extra-index-url https://pypi.nvidia.com && pip3 install --no-cache-dir https://urm.nvidia.com/artifactory/sw-ai-app-bazel-local/sw-ai-app-jarvis/tools/tensorrt_llm-0.17.0-cp312-cp312-linux_x86_64.whl pynvml==11.5.3 && pip3 install --no-cache-dir https://download.pytorch.org/whl/nightly/pytorch_triton-3.1.0%2Bcf34004b8a-cp312-cp312-linux_x86_64.whl && pip3 install --no-cache-dir nemo_toolkit['asr']==2.0.0rc1 && pip3 install --no-cache-dir huggingface-hub==0.24.0 && pip3 install --no-cache-dir pycuda==2024.1.2 && pip3 install --no-cache-dir tiktoken==0.7.0 && pip3 install --no-cache-dir transformers==4.48.0; else pip3 install --no-cache-dir torch==2.4.0; fi # buildkit
ARG TARGET_OS
ARG TARGET_ARCH
ARG TARGET_CUDA_VERSION
ENV DEBIAN_FRONTEND=noninteractive
ENV PATH=/opt/riva/bin:/opt/tritonserver/bin:/usr/local/mpi/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/ucx/bin:/opt/amazon/efa/bin
ENV LD_LIBRARY_PATH=/opt/riva/lib:/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64
ENV LD_LIBRARY_PATH=/opt/tritonserver/lib:/opt/riva/lib:/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64
ENV LD_LIBRARY_PATH=/opt/tritonserver/backends/pytorch:/opt/tritonserver/lib:/opt/riva/lib:/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64
ENV RIVA_SERVER_USER=riva-server
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c userdel $RIVA_SERVER_USER > /dev/null 2>&1 || true && if ! id -u $RIVA_SERVER_USER > /dev/null 2>&1 ; then useradd $RIVA_SERVER_USER; fi && if [ "$TARGET_OS" != "l4t" ]; then usermod -u 1002 ubuntu && groupmod -g 1002 ubuntu && usermod -u 1000 $RIVA_SERVER_USER && groupmod -g 1000 $RIVA_SERVER_USER; fi && [ `id -u $RIVA_SERVER_USER` -eq 1000 ] && [ `id -g $RIVA_SERVER_USER` -eq 1000 ] && mkdir -p /home/$RIVA_SERVER_USER/ && chmod 777 /home/$RIVA_SERVER_USER/ # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c apt-get update && apt-get install -y --no-install-recommends libb64-0d bc libcurl4-openssl-dev libexpat1 linux-libc-dev libc6 libc-bin libcurl3-gnutls libnghttp2-14 flac wget libatlas-base-dev libgoogle-glog-dev libflac++-dev libarchive-dev libogg0 libnghttp2-14 libtool locales policykit-1 python3-pip && apt-get install -y --no-install-recommends openssl libssl3 libssl-dev portaudio19-dev python3-pyaudio alsa-utils && if [ "$TARGET_OS" = "l4t" ]; then apt-get install -y libopenblas-dev libopencc-dev libopenmpi3 git curl vim unzip && cp /usr/lib/aarch64-linux-gnu/lapack/liblapack.so.3.10.0 /usr/lib/aarch64-linux-gnu/liblapack.so.3 && cp /usr/lib/aarch64-linux-gnu/blas/libblas.so.3.10.0 /usr/lib/aarch64-linux-gnu/libblas.so.3; fi && rm -rf /var/lib/apt/lists/* # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c apt-get remove -y --purge libslurm*; apt-get remove -y --purge libpmi*; apt-get remove --purge -y openssh* || true # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ -f "/opt/hpcx/ompi/etc/openmpi-mca-params.conf" ]; then echo "plm_rsh_agent = false" >> /opt/hpcx/ompi/etc/openmpi-mca-params.conf; fi # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c touch /usr/share/locale/locale.alias && sed -i '/en_US.UTF-8/s/^# //g' /etc/locale.gen && locale-gen # buildkit
WORKDIR /opt/tritonserver
COPY /opt/riva/backends/model.py /opt/riva/backends/model.py # buildkit
COPY /opt/riva/backends/megatron_model.py /opt/riva/backends/megatron_model.py # buildkit
COPY /opt/riva/backends/whisper.py /opt/riva/backends/whisper.py # buildkit
COPY /opt/riva/backends/canary.py /opt/riva/backends/canary.py # buildkit
COPY /opt/riva/backends/silero.py /opt/riva/backends/silero.py # buildkit
COPY riva/cbe/python-asr/nemo_offline_asr_model.py /opt/riva/backends/nemo_offline_asr_model.py # buildkit
COPY /opt/riva/backends/nemo_asr_model.py /opt/riva/backends/nemo_asr_model.py # buildkit
COPY /opt/riva/backends/cache_aware_model.py /opt/riva/backends/cache_aware_model.py # buildkit
COPY /opt/riva/backends/clustering_model.py /opt/riva/backends/clustering_model.py # buildkit
COPY /opt/riva/backends/nmesc_clustering.py /opt/riva/backends/nmesc_clustering.py # buildkit
COPY /opt/riva/backends/magpie_tts_model.py /opt/riva/backends/magpie_tts_model.py # buildkit
COPY /opt/riva/backends/sortformer_model.py /opt/riva/backends/sortformer_model.py # buildkit
COPY /opt/riva/backends/sortformer_modules.py /opt/riva/backends/sortformer_modules.py # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c pip install --no-cache-dir cython==3.0.11 fastapi==0.115.3 "uvicorn[standard]==0.32.0" tritonclient==${TRITON_RELEASE_VERSION} packaging==24.1 pydantic_settings==2.5.2 python-multipart==0.0.12 gunicorn==23.0.0 # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_ARCH" = "x86_64" ]; then pip install --no-cache-dir opencc==1.1.7; else pip install --no-cache-dir opencc==0.2; fi # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c pip install --no-cache-dir sacremoses==0.1.1 numpy==1.26.4 sentencepiece==0.2.0 youtokentome==1.0.6 jieba==0.42.1 pangu==4.0.6.1 # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_ARCH" = "x86_64" ]; then pip install --no-cache-dir cupy-cuda12x; elif [ "$TARGET_ARCH" = "aarch64" ]; then pip install --no-cache-dir cupy-cuda12x -f https://pip.cupy.dev/aarch64; fi # buildkit
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/sh -c if [ "$TARGET_OS" != "l4t" ]; then ln -sf ${_CUDA_COMPAT_PATH}/lib.real ${_CUDA_COMPAT_PATH}/lib && echo ${_CUDA_COMPAT_PATH}/lib > /etc/ld.so.conf.d/00-cuda-compat.conf && ldconfig && rm -f ${_CUDA_COMPAT_PATH}/lib; fi # buildkit
SHELL [/bin/bash -c]
RUN |3 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 /bin/bash -c wget -nv -O /bin/grpc_health_probe https://urm.nvidia.com/artifactory/sw-ai-app-bazel/sw-ai-app-jarvis/tools/grpc-health-probe-bca6287/grpc_health_probe_$TARGET_ARCH && chmod +x /bin/grpc_health_probe; # buildkit
SHELL [/bin/sh -c]
WORKDIR /opt/riva
ARG KALDI_DIR=submodules/kaldi
COPY /workspace/submodules/kaldi/tools/openfst-1.7.2/bin /opt/kaldi/tools/openfst-1.7.2/bin # buildkit
COPY /workspace/submodules/kaldi/tools/openfst-1.7.2/lib /opt/kaldi/tools/openfst-1.7.2/lib # buildkit
COPY /workspace/submodules/kaldi/src/fstbin /opt/kaldi/src/fstbin # buildkit
COPY /workspace/submodules/kaldi/src/lmbin /opt/kaldi/src/lmbin # buildkit
COPY /opt/riva/kenlm/build/bin/ /usr/local/bin/ # buildkit
COPY ./scripts/generate_LM_binary.sh utils/generate_LM_binary.sh # buildkit
ARG TOOLS_DIR=arpa2fst
COPY tools/asr/arpa2fst utils/arpa2fst # buildkit
ENV PATH=/opt/riva/bin:/opt/tritonserver/bin:/usr/local/mpi/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/ucx/bin:/opt/amazon/efa/bin:/opt/kaldi/tools/openfst-1.7.2/bin:/opt/kaldi/src/fstbin:/opt/kaldi/src/lmbin
ENV LD_LIBRARY_PATH=/opt/tritonserver/backends/pytorch:/opt/tritonserver/lib:/opt/riva/lib:/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64:/opt/kaldi/tools/openfst-1.7.2/lib:/opt/kaldi/src/lib
COPY VERSION / # buildkit
COPY python/servicemaker/requirements.txt /tmp/ # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c pip3 install --no-cache-dir --upgrade nvidia-pyindex==1.0.6 && pip3 install --no-cache-dir --default-timeout=100 librosa==0.10.2.post1 onnxruntime==1.17.0 onnx-graphsurgeon==0.3.27 pycuda==2024.1.2 sentencepiece==0.2.0 omegaconf==2.3.0 "sqlparse>=0.5" polygraphy==0.49.9 pytest==8.3.3 pillow==10.4.0 mock==5.1.0 deepdiff==8.0.1 python-multipart==0.0.18 setuptools==70.0.0 pypinyin==0.49.0 nltk==3.9.1 # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c if [ "$TARGET_ARCH" = "x86_64" ]; then pip3 install --no-cache-dir onnxruntime-gpu==1.17.0 --extra-index-url https://aiinfra.pkgs.visualstudio.com/PublicPackages/_packaging/onnxruntime-cuda-12/pypi/simple/; fi # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c cd /tmp && pip3 install --no-cache-dir --default-timeout=100 -r requirements.txt --extra-index-url=https://anonymous@urm.nvidia.com/artifactory/api/pypi/sw-eff-pypi-local/simple # buildkit
COPY python/servicemaker /tmp/servicemaker/ # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c cd /tmp/servicemaker/ && pip3 install --no-cache-dir .[all] # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c pip3 uninstall -y mistune # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c apt-get remove -y --purge libslurm*; apt-get remove -y --purge libpmi*; apt-get remove --purge -y openssh* || true # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c apt-get update && apt-get install --only-upgrade rsync # buildkit
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c if [ -f "/opt/hpcx/ompi/etc/openmpi-mca-params.conf" ]; then echo "plm_rsh_agent = false" >> /opt/hpcx/ompi/etc/openmpi-mca-params.conf; fi # buildkit
COPY python/download_ngc_models /usr/local/bin # buildkit
COPY python/download_and_deploy_ngc_models /usr/local/bin # buildkit
COPY python/deploy_all_models /usr/local/bin # buildkit
ENV PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION=python
ENV NUMBA_CACHE_DIR=/tmp
RUN |5 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst /bin/sh -c mkdir -p /opt/riva && tar xf /tmp/riva/riva_server_tar_distribution.tgz -C /opt/riva && tar xf /tmp/riva/riva_asrlib_tar_distribution.tgz -C /opt/riva # buildkit
ARG RIVA_NEW_TRITON_BACKENDS=riva_tts_chunker riva_tts_preprocessor riva_tts_postprocessor riva_asr_decoder riva_asr_endpointing riva_nlp_tokenizer riva_asr_features riva_nlp_qa riva_asr_lattices riva_nlp_seqlabel riva_nlp_detokenizer riva_nlp_pipeline riva_asr_pipeline riva_asr_ensemble_pipeline riva_tts_pipeline
RUN |6 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst RIVA_NEW_TRITON_BACKENDS=riva_tts_chunker riva_tts_preprocessor riva_tts_postprocessor riva_asr_decoder riva_asr_endpointing riva_nlp_tokenizer riva_asr_features riva_nlp_qa riva_asr_lattices riva_nlp_seqlabel riva_nlp_detokenizer riva_nlp_pipeline riva_asr_pipeline riva_asr_ensemble_pipeline riva_tts_pipeline /bin/sh -c for backend in ${RIVA_NEW_TRITON_BACKENDS}; do mkdir -p /opt/tritonserver/backends/${backend} && ln -s /opt/riva/backends/libtriton_${backend}.so /opt/tritonserver/backends/${backend}/; done # buildkit
ENV NVIDIA_PRODUCT_NAME=Riva Speech Server
COPY docker/entrypoint.d/ /opt/nvidia/entrypoint.d/ # buildkit
ARG RIVA_SPEECH_VERSION=devel
ARG RIVA_SPEECH_CONTAINER_VERSION=devel
ENV RIVA_SPEECH_VERSION=2.19.0
ENV NVIDIA_RIVA_SPEECH_SERVER_VERSION=25.02
LABEL com.nvidia.riva.speech.version=2.19.0
LABEL com.nvidia.riva.speech.container.version=25.02
ARG NVIDIA_BUILD_ID
ENV NVIDIA_BUILD_ID=151443007
LABEL com.nvidia.build.id=151443007
ARG NVIDIA_BUILD_REF
LABEL com.nvidia.build.ref=9fc022890a4183d9105c373028a3217dab6dddde
SHELL [/bin/bash -c]
RUN |10 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst RIVA_NEW_TRITON_BACKENDS=riva_tts_chunker riva_tts_preprocessor riva_tts_postprocessor riva_asr_decoder riva_asr_endpointing riva_nlp_tokenizer riva_asr_features riva_nlp_qa riva_asr_lattices riva_nlp_seqlabel riva_nlp_detokenizer riva_nlp_pipeline riva_asr_pipeline riva_asr_ensemble_pipeline riva_tts_pipeline RIVA_SPEECH_VERSION=2.19.0 RIVA_SPEECH_CONTAINER_VERSION=25.02 NVIDIA_BUILD_ID=151443007 NVIDIA_BUILD_REF=9fc022890a4183d9105c373028a3217dab6dddde /bin/bash -c if [ "${RIVA_SPEECH_VERSION:(-4)}" = "-dev" ] || [ "${RIVA_SPEECH_VERSION}" = "devel" ]; then github_branch="main"; else github_branch="release/${RIVA_SPEECH_VERSION}"; fi && mkdir -p /tmp/client && cd /tmp/client/ && git clone -b $github_branch https://github.com/nvidia-riva/python-clients.git && cd python-clients && python3 -m pip install --no-cache-dir -r requirements.txt && cd scripts && find . -name "*.py" | grep -v __init__.py | xargs -i cp {} /opt/riva/examples && pip install --no-cache-dir git+https://github.com/nvidia-riva/python-clients/@$github_branch && rm -rf /tmp/client/ # buildkit
SHELL [/bin/sh -c]
ENV PATH=/opt/riva/clients:/usr/local/cuda-12.8/bin:/opt/riva/bin:/opt/tritonserver/bin:/usr/local/mpi/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/ucx/bin:/opt/amazon/efa/bin:/opt/kaldi/tools/openfst-1.7.2/bin:/opt/kaldi/src/fstbin:/opt/kaldi/src/lmbin
ENV LD_LIBRARY_PATH=/usr/local/cuda-12.8/lib64:/opt/tritonserver/backends/pytorch:/opt/tritonserver/lib:/opt/riva/lib:/usr/local/cuda/compat/lib:/usr/local/nvidia/lib:/usr/local/nvidia/lib64:/opt/kaldi/tools/openfst-1.7.2/lib:/opt/kaldi/src/lib
COPY scripts/calc_wer.py utils/calc_wer.py # buildkit
COPY test_files/asr/public wav # buildkit
COPY test_files/asr/wav/test/1272-135031-0000.alaw /opt/riva/examples/asr_lib/ # buildkit
COPY test_files/asr/wav/test/1272-135031-0000.mulaw /opt/riva/examples/asr_lib/ # buildkit
COPY test_files/asr/wav/test/1272-135031-0000.wav /opt/riva/examples/asr_lib/ # buildkit
RUN |10 TARGET_OS=linux TARGET_ARCH=x86_64 TARGET_CUDA_VERSION=12.8 KALDI_DIR=submodules/kaldi TOOLS_DIR=arpa2fst RIVA_NEW_TRITON_BACKENDS=riva_tts_chunker riva_tts_preprocessor riva_tts_postprocessor riva_asr_decoder riva_asr_endpointing riva_nlp_tokenizer riva_asr_features riva_nlp_qa riva_asr_lattices riva_nlp_seqlabel riva_nlp_detokenizer riva_nlp_pipeline riva_asr_pipeline riva_asr_ensemble_pipeline riva_tts_pipeline RIVA_SPEECH_VERSION=2.19.0 RIVA_SPEECH_CONTAINER_VERSION=25.02 NVIDIA_BUILD_ID=151443007 NVIDIA_BUILD_REF=9fc022890a4183d9105c373028a3217dab6dddde /bin/sh -c mkdir /data/ && chmod 777 /data/ # buildkit

Labels

Key Value
com.nvidia.build.id 151443007
com.nvidia.build.ref 9fc022890a4183d9105c373028a3217dab6dddde
com.nvidia.cublas.version 12.8.3.14
com.nvidia.cuda.version 9.0
com.nvidia.cudnn.version 9.7.0.66
com.nvidia.cufft.version 11.3.3.41
com.nvidia.curand.version 10.3.9.55
com.nvidia.cusolver.version 11.7.2.55
com.nvidia.cusparse.version 12.5.7.53
com.nvidia.cusparselt.version 0.6.3.2
com.nvidia.cutensor.version 2.1.0.9
com.nvidia.nccl.version 2.25.1
com.nvidia.npp.version 12.3.3.65
com.nvidia.nsightcompute.version 2025.1.0.14
com.nvidia.nsightsystems.version 2024.6.2.225
com.nvidia.nvjpeg.version 12.3.5.57
com.nvidia.riva.speech.container.version 25.02
com.nvidia.riva.speech.version 2.19.0
com.nvidia.tensorrt.version 10.8.0.43
com.nvidia.tensorrtoss.version
com.nvidia.volumes.needed nvidia_driver
org.opencontainers.image.ref.name ubuntu
org.opencontainers.image.version 24.04
Details
Container
2025-07-02 05:09:06 +00:00
11
OCI / Docker
linux/amd64
20 GiB
Versions (2) View all
2.19.0 2025-07-02
latest 2025-07-01