I am Starting Deepstream pipelines with Deepstream Dockers. but sometimes I am getting this error
``Error parsing GPU utilization: Failed to initialize NVML: Unknown Error\n[2025-09-24 07:37:02] âï¸ nvidia-smi failed with exit code: 65280```
Allright, I installed the correct Driver. But I am getting this.
Error parsing GPU utilization: Failed to initialize NVML: Unknown Error
[2025-09-26 16:44:48] �~Z| �~O nvidia-smi failed with exit code: 65280
did you run “nvidia-smi” in host? if so, it seems that driver is not installed correctly. Please uninstall the old driver first, then install the new one, then reboot. Here is the guide to use DeepStream docker.
nvidia-smi works correctly
but still this issue comes up
and it is random. sometimes comes up when I spawn docker container or sometimes hours after the docker has been running
What is your docker start command? In docker container, can deepstream-test1 run well? which DeepStream pipeline causes the error “Error parsing GPU utilization: Failed to initialize NVML”?
Above is my Docker Compose file. Below is my docker file.
ARG VERSION
FROM ha-edge-base:$VERSION AS dgpu-base
WORKDIR /app
# To get video driver libraries at runtime (libnvidia-encode.so/libnvcuvid.so)
ENV NVIDIA_DRIVER_CAPABILITIES=$NVIDIA_DRIVER_CAPABILITIES,video,compute,graphics,utility
ENV CUDA_HOME=/usr/local/cuda
ENV CFLAGS="-I$CUDA_HOME/include $CFLAGS"
ENV RUNNING_IN_DOCKER=true
ENV LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:/usr/local/lib
# Fix for Ultralytics YOLO config directory warning
ENV YOLO_CONFIG_DIR=/tmp
# Set alias for faster builds
RUN echo 'alias build-app="meson setup runtime/build --reconfigure && ninja -C runtime/build install -j12"' >> ~/.bashrc
To narrow down this issue, If starting the DeepStream 7.1 docker with the method in the guide, can “nvidia-smi” and deepstream-test1 run well in docker container?
About “Failed to initialize NVML”, Please refer to this topic.
[1] Yes it does. I can see output of nvidia-smi and can run deepstream-test1
[2] Okay.
The issue is, It works sometimes and it does not work the other times. At the moment I do not know under what conditions it fails.
when It fails, We get error in this default print generated by Deepstream Docker.
===============================
hawk-edge-09 | DeepStreamSDK
hawk-edge-09 | ===============================
hawk-edge-09 |
hawk-edge-09 | *** LICENSE AGREEMENT ***
hawk-edge-09 | By using this software you agree to fully comply with the terms and conditions
hawk-edge-09 | of the License Agreement. The License Agreement is located at
hawk-edge-09 | /opt/nvidia/deepstream/deepstream/LicenseAgreement.pdf. If you do not agree
hawk-edge-09 | to the terms and conditions of the License Agreement do not use the software.
hawk-edge-09 |
hawk-edge-09 |
hawk-edge-09 | =============================
hawk-edge-09 | == Triton Inference Server ==
hawk-edge-09 | =============================
hawk-edge-09 |
hawk-edge-09 | NVIDIA Release 24.08 (build 107631419)
hawk-edge-09 | Triton Server Version 2.49.0
hawk-edge-09 |
hawk-edge-09 | Copyright (c) 2018-2024, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
hawk-edge-09 |
hawk-edge-09 | Various files include modifications (c) NVIDIA CORPORATION & AFFILIATES. All rights reserved.
hawk-edge-09 |
hawk-edge-09 | This container image and its contents are governed by the NVIDIA Deep Learning Container License.
hawk-edge-09 | By pulling and using the container, you accept the terms and conditions of this license:
hawk-edge-09 | https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license
hawk-edge-09 |
From the log in your last comment, there was no any error printing. there was only some logs of Triton starting. Do you mean if starting the docker with method in the guide, nvidia-smi and deepstream-test1 work well every time? if so, please simplify the custom docker start method to narrow down this issue. Here is a sample.