Hi! When I serialize a particular network and then deserialize it, it crashes somewhere inside the nvinfer.dll 4 levels deep without a callstack. It happens with one particular network, not with all, it’s huge (~2.1Gb) and contains custom layers (which are implemented).
On Windows there’s no info, but on Ubuntu I get traces:
trt: runtime.cpp (24) - Cuda Error in allocate: 2
trt: cuda/cudaFusedConvActLayer.cpp (287) - Cuda Error in executeFused: 2
trt: cuda/cudaFusedConvActLayer.cpp (287) - Cuda Error in executeFused: 2
Could you please give advice on what might be happening? Thanks!
Reproduced in environments:
OS: Windows 10 x64, TensorRT 5.1.5.0 (for CUDA 10.1), CUDA 10.1.105, MSVC 2017, GPU GTX 1060.
OS: Ubuntu 16.04
Hi, can you provide the following details on the platforms you are using?
Linux distro and version
Ubuntu 16.04
GPU type
GTX 1060
Nvidia driver version
CUDA version
CUDNN version
Python version [if using python]
TensorFlow version
TensorRT version
Any source files and models you can provide will help us reproduce your issue and further debug it. You can private message these if you don’t want them to be public.