Hi,
Thanks for your question.
You can train a model with DIGITs and deploy it on TX2 with tensorRT.
DIGITs will save model with default caffe format which is independent to GPU architecture.
Once you load the model with tensorRT, it will re-compile the model based on TX2 GPU architecture.
More, if your input is video, usually we divide it into frames and run prediction for each frame separately.
Here is more information:
https://developer.nvidia.com/embedded/twodaystoademo