I had trained a ResNet50 model using my own dataset for image classification via NVIDIA TLT. The output of the toolkit is resnet50.etlt and resnet50.trt. I’m trying to export the model to NVIDIA triton server, however, the server requires a model.plan format file and a config file.
How can I deploy my model to triton server smoothly, it seems that I have to do some kind of work to transfer my resnet50.trt into model.plan for the server.
I’m using
TLT docker image: nvcr.io/nvidia/tlt-streamanalytics:v3.0-dp-py3 (TRT:7.2.1)
Triton docker image: nvcr.io/nvidia/tritonserver:20.12-py3 (TRT:7.2.2)
System:
Ubuntu 16.04
NVIDIA Tesla V100