• Hardware Platform (Jetson / GPU) -Jetson Orin
• DeepStream Version - 6.4
• JetPack Version (valid for Jetson only) - 6.0+b106
• TensorRT Version - 8.6.2
• NVIDIA GPU Driver Version (valid for GPU only) - 12.2
• Issue Type( questions, new requirements, bugs) - we are try to deploy our model into triton server but getting an error unable to load model and peoplenet_resnet32 model is used so kindly help me to resolve the problem
give right way to implement triton server on jetson orin with cuda
• How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing)
• Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description)
jetson@ubuntu:~/server$ sudo docker run --rm --runtime=nvidia --net=host --gpus all -v /home/jetson/server/model_repository:/models nvcr.io/nvidia/tritonserver:23.10-py3 tritonserver --model-repository=/models
=============================
== Triton Inference Server ==
NVIDIA Release 23.10 (build 72127510)
Triton Server Version 2.39.0
Copyright (c) 2018-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
Various files include modifications (c) NVIDIA CORPORATION & AFFILIATES. All rights reserved.
This container image and its contents are governed by the NVIDIA Deep Learning Container License.
By pulling and using the container, you accept the terms and conditions of this license:
Failed to detect NVIDIA driver version.
I0628 12:48:49.197002 1 pinned_memory_manager.cc:241] Pinned memory pool is created at ‘0x203eae000’ with size 268435456
I0628 12:48:49.197413 1 cuda_memory_manager.cc:107] CUDA memory pool is created on device 0 with size 67108864
E0628 12:48:49.203240 1 model_repository_manager.cc:1309] Poll failed for model directory ‘models’: Invalid model name: Could not determine backend for model ‘models’ with no backend in model configuration. Expected model name of the form ‘model.<backend_name>’.
I0628 12:48:49.205591 1 model_lifecycle.cc:461] loading: peoplenet:1
I0628 12:48:49.289920 1 tensorrt.cc:65] TRITONBACKEND_Initialize: tensorrt
I0628 12:48:49.289977 1 tensorrt.cc:75] Triton TRITONBACKEND API version: 1.16
I0628 12:48:49.289985 1 tensorrt.cc:81] ‘tensorrt’ TRITONBACKEND API version: 1.16
I0628 12:48:49.289992 1 tensorrt.cc:105] backend configuration:
{“cmdline”:{“auto-complete-config”:“true”,“backend-directory”:“/opt/tritonserver/backends”,“min-compute-capability”:“6.000000”,“default-max-batch-size”:“4”}}
I0628 12:48:49.290495 1 tensorrt.cc:231] TRITONBACKEND_ModelInitialize: peoplenet (version 1)
I0628 12:48:49.333198 1 logging.cc:46] Loaded engine size: 22 MiB
E0628 12:48:49.340913 1 logging.cc:40] 6: The engine plan file is not compatible with this version of TensorRT, expecting library version 8.6.1.6 got 8.6.2.3, please rebuild.
E0628 12:48:49.359902 1 logging.cc:40] 2: [engine.cpp::deserializeEngine::951] Error Code 2: Internal Error (Assertion engine->deserialize(start, size, allocator, runtime) failed. )
I0628 12:48:49.364211 1 tensorrt.cc:274] TRITONBACKEND_ModelFinalize: delete model state
E0628 12:48:49.364280 1 model_lifecycle.cc:621] failed to load ‘peoplenet’ version 1: Internal: unable to load plan file to auto complete config: /models/peoplenet/1/model.engine
I0628 12:48:49.364306 1 model_lifecycle.cc:756] failed to load ‘peoplenet’
I0628 12:48:49.364469 1 server.cc:592]
±-----------------±-----+
| Repository Agent | Path |
±-----------------±-----+
±-----------------±-----+
I0628 12:48:49.364541 1 server.cc:619]
±---------±----------------------------------------------------------±--------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Backend | Path | Config |
±---------±----------------------------------------------------------±--------------------------------------------------------------------------------------------------------------------------------------------------------------+
| tensorrt | /opt/tritonserver/backends/tensorrt/libtriton_tensorrt.so | {“cmdline”:{“auto-complete-config”:“true”,“backend-directory”:“/opt/tritonserver/backends”,“min-compute-capability”:“6.000000”,“default-max-batch-size”:“4”}} |
±---------±----------------------------------------------------------±--------------------------------------------------------------------------------------------------------------------------------------------------------------+
I0628 12:48:49.364585 1 server.cc:662]
±----------±--------±----------------------------------------------------------------------------------------------------------+
| Model | Version | Status |
±----------±--------±----------------------------------------------------------------------------------------------------------+
| peoplenet | 1 | UNAVAILABLE: Internal: unable to load plan file to auto complete config: /models/peoplenet/1/model.engine |
±----------±--------±----------------------------------------------------------------------------------------------------------+
Driver is unsupported. Must be at least 384.00.
W0628 12:48:49.375665 1 metrics.cc:738] DCGM unable to start: DCGM initialization error
I0628 12:48:49.376266 1 metrics.cc:710] Collecting CPU metrics
I0628 12:48:49.376567 1 tritonserver.cc:2458]
±---------------------------------±----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Option | Value |
±---------------------------------±----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| server_id | triton |
| server_version | 2.39.0 |
| server_extensions | classification sequence model_repository model_repository(unload_dependents) schedule_policy model_configuration system_shared_memory cuda_shared_memory binary_tensor_data parameters statistics trace logging |
| model_repository_path[0] | /models |
| model_control_mode | MODE_NONE |
| strict_model_config | 0 |
| rate_limit | OFF |
| pinned_memory_pool_byte_size | 268435456 |
| cuda_memory_pool_byte_size{0} | 67108864 |
| min_supported_compute_capability | 6.0 |
| strict_readiness | 1 |
| exit_timeout | 30 |
| cache_enabled | 0 |
±---------------------------------±----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
I0628 12:48:49.376589 1 server.cc:293] Waiting for in-flight requests to complete.
I0628 12:48:49.376600 1 server.cc:309] Timeout 30: Found 0 model versions that have in-flight inferences
I0628 12:48:49.376609 1 server.cc:324] All models are stopped, unloading models
I0628 12:48:49.376615 1 server.cc:331] Timeout 30: Found 0 live models and 0 in-flight non-inference requests
error: creating server: Internal - failed to load all models
jetson@ubuntu:~/server$ nvidia-smi
Fri Jun 28 18:21:33 2024
±--------------------------------------------------------------------------------------+
| NVIDIA-SMI 540.3.0 Driver Version: N/A CUDA Version: 12.2 |
|-----------------------------------------±---------------------±---------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 Orin (nvgpu) N/A | N/A N/A | N/A |
| N/A N/A N/A N/A / N/A | Not Supported | N/A N/A |
| | | N/A |
±----------------------------------------±---------------------±---------------------+
±--------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| No running processes found |
±--------------------------------------------------------------------------------------+