Hi,
Thanks for sharing.
[05/08/2024-10:14:17] [E] Error[10]: [optimizer.cpp::computeCosts::3728] Error Code 10: Internal Error (Could not find any implementation for node {ForeignNode[/cnn/cnn.0/Conv]}.)
[05/08/2024-10:14:17] [V] [TRT] =============== Computing costs for
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(6144,2048,64,1) -> Int8(65536,1024,64,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(6144,2048,64,1) -> Int8(512,256:32,16,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(1024,1:4,32,1) -> Int8(65536,1024,64,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(1024,1:4,32,1) -> Int8(512,256:32,16,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(1024,1024:32,32,1) -> Int8(65536,1024,64,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
[05/08/2024-10:14:17] [V] [TRT] *************** Autotuning format combination: Int8(1024,1024:32,32,1) -> Int8(512,256:32,16,1) ***************
[05/08/2024-10:14:17] [V] [TRT] --------------- Timing Runner: {ForeignNode[/cnn/cnn.0/Conv]} (DLA)
[05/08/2024-10:14:17] [V] [TRT] Setting a default quantization params because quantization data is missing for {ForeignNode[/cnn/cnn.0/Conv]}
[05/08/2024-10:14:17] [W] [TRT] Skipping tactic 0x0000000000000003 due to exception Failed to create DLA runtime context. Hint: You can load at most 16 DLA loadables simultaneously per core. Attempting to load more will cause context allocation to fail.
[05/08/2024-10:14:17] [V] [TRT] Fastest Tactic: 0xd15ea5edd15ea5ed Time: inf
Based on the log, the model already meets the DLA loadable limit.
In such case, TensorRT should fall back to GPU for inference but it doesn’t.
Could you test it on JetPack 5.1.3? The latest software for XavierNX?
Thanks.