Hey there,
I’m having a hard time running Clara Deploy SDK and got stuck at “Wait until TRTIS is Ready” as it’s mentioned below while running as per the following https://ngc.nvidia.com/catalog/containers/nvidia:clara:ai-covid-19 link. And I kindly request you to look into it and let me know the resolution and what the problem via the content that is as follows:
ubuntu@ip-172-31-22-181:~$ ./run_docker.sh
Error response from daemon: network with name container-demo already exists
0538adc89e9e0961808bd6c687bb5fa79035ab9c7bf7e812ec5742acc99da427
nvidia-docker run --name trtis --network container-demo -d --rm --shm-size=1g --ulimi
t memlock=-1 --ulimit stack=67108864 -p 8000:8000 -v /home/ubuntu/models/classificati
on_covid-19_v1:/models/classification_covid-19_v1 nvcr.io/nvidia/tensorrtserver:19.08
-py3 trtserver --model-store=/models
Wait until TRTIS 172.18.0.2 is ready…
You may observe in the above-posted screenshots that it got stuck while executing the following shell script at https://ngc.nvidia.com/catalog/containers/nvidia:clara:ai-covid-19
./run_docker.sh
I did the above one soon after setting up with the reference pipeline using the following link https://ngc.nvidia.com/catalog/model-scripts/nvidia:clara:clara_ai_covid19_pipeline
Could you at least let me know where exactly I went wrong? Or Did I do anything in a different order? Or Didn’t store the files in the appropriate directories? I’m kinda confused here!
Or Do I have to do anything with the NVIDIA Triton Inference Server (Triton), formerly known as TensorRT Inference Server (TRTIS)?
And I was also following the Clara Deploy SDK Docs https://docs.nvidia.com/clara/deploy/index.html page while doing the above ones.
But I guess but I don’t know exactly if I’m correct in this regard that the problem is with the NVIDIA Triton Inference Server (Triton), formerly known as TensorRT Inference Server (TRTIS). Kindly please look into it and let me know ASAP.