I have used TensorRT’s sample uff Mask RCNN example to convert our model from h5 and create an engine file. I have loaded this engine onto the Triton server and used the --strict-model-config=false flag to generate the config.pbtxt. The server is started successfully on the HTTP and gRPC ports. What I need help with is using the Triton C++ API to write the inference code. Not sure which client examples to follow. Any lead or direction of how to go about this or which particular example to follow would really help! Thank you.
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
Mask RCNN TensorRT in Triton | 3 | 575 | July 9, 2020 | |
Looking for examples of inference using C++ | 2 | 360 | October 18, 2021 | |
Triton Inference Server | 1 | 423 | June 21, 2023 | |
Initialise Triton data with inputs | 0 | 610 | August 10, 2020 | |
Inference server - resnet50 int8 issue | 1 | 835 | September 3, 2019 | |
Deserialize tensorrt model C++ API | 0 | 383 | March 30, 2021 | |
TensorRT .trt custom model inference | 2 | 446 | December 4, 2020 | |
Tensorrt inference | 0 | 415 | March 30, 2022 | |
TensorRT deployment with engine generated from TLT example | 8 | 776 | December 5, 2020 | |
Want example about performing action recognition with tensorrt | 4 | 629 | April 8, 2021 |