• Hardware: Jetson Nano
• Network Type: Faster RCNN
• TLT Version 3.0
Hi, I’m trying to deploy a Faster RCNN model (trained in a x86 machine) on DeepStream. But when I execute the tlt-converter
command on the Jetson Nano, it responds with Cuda Error.
Command: ./tlt-converter -k NnRkcW8zdWk4ZG1hMWpvdW44OGpuODA1azc6ZDZiNDVkMjQtYzJlYS00OTgzLWEzNTUtNmIxMWE4NDc3Nzk5 -d 3,480,640 -o NMS frcnn_kitti_mobilenet_v1_train_fp32.etlt
Response log: [INFO] Some tactics do not have sufficient workspace memory to run. Increasing workspace size may increase performance, please check verbose output. [ERROR] ../builder/cudnnBuilderUtils.cpp (427) - Cuda Error in findFastestTactic: 702 (the launch timed out and was terminated) [ERROR] ../rtSafe/safeRuntime.cpp (32) - Cuda Error in free: 702 (the launch timed out and was terminated) terminate called after throwing an instance of 'nvinfer1::CudaError' what(): std::exception Aborted (core dumped)
I tried increasing the workspace using the -w flag, but the response is the same.
Can you help me please?
May I know more info about the environment?
What is the CUDA/Cudnn/TensorRT version? Did you install them via Jetpack?
You can run $dpkg -l |grep cuda
I got the same response as before
[INFO] Some tactics do not have sufficient workspace memory to run. Increasing workspace size may increase performance, please check verbose output. [ERROR] ../builder/cudnnBuilderUtils.cpp (427) - Cuda Error in findFastestTactic: 702 (the launch timed out and was terminated) [ERROR] ../rtSafe/safeRuntime.cpp (32) - Cuda Error in free: 702 (the launch timed out and was terminated) terminate called after throwing an instance of 'nvinfer1::CudaError' what(): std::exception Aborted (core dumped)
Please try other models.
See Tlt-converter failed on Jetson Nano with Cuda Error in loadKernel: 702
It has similar error.
But actually I cannot reproduce. And the error results from end user.
Trying with Yolo v3 and Yolo v4 models responds with key error, but using SSD and RetinaNet generates engine successfully.
[ERROR] UffParser: Unsupported number of graph 0
[ERROR] Failed to parse the model, please check the encoding key to make sure it's correct
[ERROR] Network must have at least one output
[ERROR] Network validation failed.
[ERROR] Unable to create engine
Segmentation fault (core dumped)
What is the command when you run Yolo v3 and Yolo v4 ?
./tlt-converter -k nvidia_tlt -d 3,544,960 -o BatchedNMS yolov3_resnet18.etlt
./tlt-converter -k nvidia_tlt -d 3,544,960 -o BatchedNMS yolov4_resnet18.etlt
Please run with below.
$ ./tlt-converter -k nvidia_tlt -d 3,544,960 -p Input,1x3x544x960,1x3x544x960,2x3x544x960 yolov3_resnet18.etlt
I got the following error:
./tlt-converter: invalid option -- '-'
Unrecognized argument
Aborted (core dumped)
How did you download tlt-converter?
From Deploying to DeepStream / Generating an Engine Using tlt-converter / Instructions for Jetson , on the Faster RCNN TLT documentation.
Here is the link:
Faster RCNN TLT
Can you share the result of “$ ./tlt-converter -h” ?