I am porting a C++ program from the P4 platform to the T4 platform, which is work fine on P4. But on T4 could not, and the error messages are related to illegal memory. Dos T4 not support running multiple algorithm models on a single GPU card?
Can you please provide a small repro package with code and other necessary data so I can further debug this? It is possible that code from TensorRT 3 has since been deprecated in TensorRT 5.
I have some models in Caffe and then optimized by TensorRT. They are running well in the same thread but I am having some problems when using mulit-threads,each Thread run a model. I am pasting the error just below:
ERROR: CUDA cask failure at execution for trt_maxwell_scudnn_128x32_relu_small_nn_v1.
ERROR: cuda/caskConvolutionLayer.cpp (256) - Cuda Error in execute: 77
ERROR: cuda/caskConvolutionLayer.cpp (256) - Cuda Error in execute: 77
I’ve generated the PLAN file from a turing uarchitecture (GeForce 1050 Ti and T4) so it sounds weird for me that part that sais “failure at execution for trt_maxwell_scudnn_128x32_relu_small_nn_v1”. Does it have any sense?
1. original model
2. converted model
3. scripts that you ran to convert the model
4. scripts for running the model with multiple threads
5. dataset used for models/scripts if any
so I can reproduce and debug this issue? You can private message this information to me if you don’t want to share it publicly.