Hi,
There is two source consumes memory:
1. Loading libraries: (TensorRT, cuDNN, cuBLAS…)
- Amount: around 600Mib (TensorRT3)
- Required but is shared with all the processes.
2. Building inference engine:
- Amount: depends on the network size
- Can be limited by setMaxWorkspaceSize() and setMaxBatchSize(). Each process has their own consumption.
Thanks.