Unable to generate tensorrt engine for facenet model on Jetson Nano 2 GB

Scenario A

Tried to allocate swap memory of 6GB through the script:

As per the “swap_area_modified_killed.png” snap, the process gets killed after using 1915MB RAM area out of 1972MB total RAM and 2017MB of SWAP area from 7144MB(Total ~4GB from ~9GB)


UNKNOWN: Retargeting InceptionResnetV1/Repeat_2/block8_5/Branch_1/Conv2d_0c_3x1/Relu to InceptionResnetV1/Repeat_2/block8_5/concat
UNKNOWN: Eliminating concatenation InceptionResnetV1/Block8/concat
UNKNOWN: Generating copy for InceptionResnetV1/Block8/Branch_1/Conv2d_0a_1x1/convolution + InceptionResnetV1/Block8/Branch_1/Conv2d_0a_1x1/Relu || InceptionResnetV1/Block8/Branch_0/Conv2d_1x1/convolution + InceptionResnetV1/Block8/Branch_0/Conv2d_1x1/Relu to InceptionResnetV1/Block8/concat
UNKNOWN: Retargeting InceptionResnetV1/Block8/Branch_1/Conv2d_0c_3x1/Relu to InceptionResnetV1/Block8/concat
UNKNOWN: After concat removal: 137 layers
UNKNOWN: Graph construction and optimization completed in 1.2832 seconds.
Killed

nano@nano-desktop:~$ swapon -s
Filename Type Size Used Priority
/mnt/swapfile file 6291452 0 -1
/dev/zram0 partition 255996 0 5
/dev/zram1 partition 255996 0 5
/dev/zram2 partition 255996 0 5
/dev/zram3 partition 255996 0 5

Scenario B

Tried to allocate swap area 1GB to each CPU ZRAM:

[zram_modified_killed|690x387]
(https://forums.developer.nvidia.com/uploads/short-url/pv2TwAN8xu1FMGFMN4ch6uPNQIK.jpeg)

As per the “zram_modified_killed.png” snap, the process gets killed after using 1915MB RAM area out of 1972MB total RAM and 2017MB of SWAP area from 7144MB(Total ~4GB from ~9GB)

$ sudo tegrastats
RAM 1904/1972MB (lfb 1x1MB) SWAP 2030/40000MB (cached 0MB) IRAM 0/252kB(lfb 252kB)

nano@nano-desktop:~$ swapon -s
Filename Type Size Used Priority
/dev/zram0 partition 1023996 7056 5
/dev/zram1 partition 1023996 7052 5
/dev/zram2 partition 1023996 7068 5
/dev/zram3 partition 1023996 7064 5

In the both the scenarios, the process got killed after using ~4GB(2GB swap & 2GB RAM) even the swap area is still not fully occupied.

Hi,

It looks like your application is a deep-learning use case.

Please noted that GPU could not use the swap memory.
So to increase swap memory won’t have too much help for CUDA inference.

Thanks.