TensorRT and Triton Server - different results each time

We’re running a YOLOv5s model converted to a TensorRT engine in a Triton Server. Sending the same image to the Triton server for inference returns different results each time. Is this normal or is there something we could do to make it deterministic? We are using a Jetson Nano 2GB and built the TensorRT engine on the Jetson Nano itself.

1 Like