Although it isn’t jetson_inference, you can find an example with TensorRT in the below comment:
The example serializes a TensorRT engine and deploys it to given hardware (GPU or DLA).
When input more than one hardware, there are multiple engines loaded and running in threads.