Deepstream Triton Docker: nvcr.io/nvidia/deepstream:5.0.1-20.09-triton
Serving Triton Models Repository : trtserver --model-repository=/models/ --log-verbose=1 --strict-model-config=False
Note : In Triton server the command to serve models is
tritonserver --model-repository=/models/ --strict-model-config=False --log-verbose=1 --strict-model-config=False
and in deepstream its “trtserver”
Issue : HTTP error when trying to access localhost docker apis from triton-perf container
APIs loaded and running Logs :
I1102 07:27:58.614903 522 grpc_server.cc:1939] Started GRPCService at 0.0.0.0:8001
I1102 07:27:58.614922 522 http_server.cc:1411] Starting HTTPService at 0.0.0.0:8000
I1102 07:27:58.656984 522 http_server.cc:1426] Starting Metrics Service at 0.0.0.0:8002
Trying to access Model API Endpoints:
Command : perf_client -u localhost:8000 -m model_name -b 1 --concurrency 1
Error Output : error: failed to get model metadata: failed to parse the request JSON buffer: The document is empty. at 0
When I check the triton Logs, I get these outputs:
I1102 07:28:12.073094 522 http_server.cc:456] HTTP request: 0 /v2/models/model_name
I1102 07:28:12.073159 522 http_server.cc:509] HTTP error: 0 /v2/models/model_name - 400
I don’t have these issues when I use triton server. Why is this issue arising with Deepstream-Triton? How can I resolve this?
Thanks.