The following is my operation record. Please help me see if there is any problem?
xiayd@xiayd-MS-7C82:~$ docker login nvcr.io
Authenticating with existing credentials...
WARNING! Your password will be stored unencrypted in /home/xiayd/.docker/config.json.
Configure a credential helper to remove this warning. See
https://docs.docker.com/engine/reference/commandline/login/#credentials-store
Login Succeeded
xiayd@xiayd-MS-7C82:~$ docker run --runtime=nvidia -it -v /home/xiayd/:/workspace/tlt-experiments nvcr.io/nvidia/tlt-streamanalytics:v3.0-py3 /bin/bash
--2021-08-20 12:25:01-- https://ngc.nvidia.com/downloads/ngccli_reg_linux.zip
Resolving ngc.nvidia.com (ngc.nvidia.com)... 143.204.128.103, 143.204.128.57, 143.204.128.11, ...
Connecting to ngc.nvidia.com (ngc.nvidia.com)|143.204.128.103|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 25071865 (24M) [application/zip]
Saving to: ‘/opt/ngccli/ngccli_reg_linux.zip’
ngccli_reg_linux.zi 100%[===================>] 23.91M 6.04MB/s in 4.0s
2021-08-20 12:25:11 (6.04 MB/s) - ‘/opt/ngccli/ngccli_reg_linux.zip’ saved [25071865/25071865]
Archive: /opt/ngccli/ngccli_reg_linux.zip
inflating: /opt/ngccli/ngc
extracting: /opt/ngccli/ngc.md5
root@f3250a75bdd4:/workspace# ls
EULA.pdf README.md tlt-experiments
root@f3250a75bdd4:/workspace# ls tlt-experiments/faster_rcnn/specs
default_spec_darknet19.txt default_spec_resnet18.txt
default_spec_darknet53.txt default_spec_resnet18_grayscale.txt
default_spec_efficientnet_b0.txt default_spec_resnet18_retrain_spec.txt
default_spec_efficientnet_b1.txt default_spec_resnet34.txt
default_spec_googlenet.txt default_spec_resnet50.txt
default_spec_mobilenet_v1.txt default_spec_vgg16.txt
default_spec_mobilenet_v2.txt default_spec_vgg19.txt
default_spec_resnet10.txt frcnn_tfrecords_kitti_trainval.txt
default_spec_resnet101.txt
root@f3250a75bdd4:/workspace# faster_rcnn export --gpu_index 0
-m tlt-experiments/faster_rcnn/frcnn_kitti_resnet18_retrain.epoch12.tlt
-o tlt-experiments/faster_rcnn/frcnn_kitti_resnet18_retrain_int8.etlt
-e tlt-experiments/faster_rcnn/specs/default_spec_resnet18_retrain_spec.txt
-k tlt
--data_type int8
--batch_size 8
--batches 10
--cal_cache_file tlt-experiments/faster_rcnn/cal.bin
Using TensorFlow backend.
Using TensorFlow backend.
WARNING:tensorflow:Deprecation warnings have been disabled. Set TF_ENABLE_DEPRECATION_WARNINGS=1 to re-enable them.
2021-08-20 12:28:02,748 [INFO] iva.common.export.keras_exporter: Using input nodes: ['input_image']
2021-08-20 12:28:02,748 [INFO] iva.common.export.keras_exporter: Using output nodes: ['NMS']
2021-08-20 12:28:02,748 [INFO] iva.faster_rcnn.spec_loader.spec_loader: Loading experiment spec at tlt-experiments/faster_rcnn/specs/default_spec_resnet18_retrain_spec.txt.
__________________________________________________________________________________________________
Layer (type) Output Shape Param # Connected to
==================================================================================================
input_image (InputLayer) (None, 3, 384, 1248) 0
__________________________________________________________________________________________________
conv1 (Conv2D) (None, 48, 192, 624) 7056 input_image[0][0]
__________________________________________________________________________________________________
bn_conv1 (BatchNormalization) (None, 48, 192, 624) 192 conv1[0][0]
__________________________________________________________________________________________________
activation_1 (Activation) (None, 48, 192, 624) 0 bn_conv1[0][0]
__________________________________________________________________________________________________
block_1a_conv_1 (Conv2D) (None, 64, 96, 312) 27648 activation_1[0][0]
__________________________________________________________________________________________________
block_1a_bn_1 (BatchNormalizati (None, 64, 96, 312) 256 block_1a_conv_1[0][0]
__________________________________________________________________________________________________
block_1a_relu_1 (Activation) (None, 64, 96, 312) 0 block_1a_bn_1[0][0]
__________________________________________________________________________________________________
block_1a_conv_2 (Conv2D) (None, 64, 96, 312) 36864 block_1a_relu_1[0][0]
__________________________________________________________________________________________________
block_1a_conv_shortcut (Conv2D) (None, 64, 96, 312) 3072 activation_1[0][0]
__________________________________________________________________________________________________
block_1a_bn_2 (BatchNormalizati (None, 64, 96, 312) 256 block_1a_conv_2[0][0]
__________________________________________________________________________________________________
block_1a_bn_shortcut (BatchNorm (None, 64, 96, 312) 256 block_1a_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_1 (Add) (None, 64, 96, 312) 0 block_1a_bn_2[0][0]
block_1a_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_1a_relu (Activation) (None, 64, 96, 312) 0 add_1[0][0]
__________________________________________________________________________________________________
block_1b_conv_1 (Conv2D) (None, 64, 96, 312) 36864 block_1a_relu[0][0]
__________________________________________________________________________________________________
block_1b_bn_1 (BatchNormalizati (None, 64, 96, 312) 256 block_1b_conv_1[0][0]
__________________________________________________________________________________________________
block_1b_relu_1 (Activation) (None, 64, 96, 312) 0 block_1b_bn_1[0][0]
__________________________________________________________________________________________________
block_1b_conv_2 (Conv2D) (None, 64, 96, 312) 36864 block_1b_relu_1[0][0]
__________________________________________________________________________________________________
block_1b_conv_shortcut (Conv2D) (None, 64, 96, 312) 4096 block_1a_relu[0][0]
__________________________________________________________________________________________________
block_1b_bn_2 (BatchNormalizati (None, 64, 96, 312) 256 block_1b_conv_2[0][0]
__________________________________________________________________________________________________
block_1b_bn_shortcut (BatchNorm (None, 64, 96, 312) 256 block_1b_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_2 (Add) (None, 64, 96, 312) 0 block_1b_bn_2[0][0]
block_1b_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_1b_relu (Activation) (None, 64, 96, 312) 0 add_2[0][0]
__________________________________________________________________________________________________
block_2a_conv_1 (Conv2D) (None, 128, 48, 156) 73728 block_1b_relu[0][0]
__________________________________________________________________________________________________
block_2a_bn_1 (BatchNormalizati (None, 128, 48, 156) 512 block_2a_conv_1[0][0]
__________________________________________________________________________________________________
block_2a_relu_1 (Activation) (None, 128, 48, 156) 0 block_2a_bn_1[0][0]
__________________________________________________________________________________________________
block_2a_conv_2 (Conv2D) (None, 128, 48, 156) 147456 block_2a_relu_1[0][0]
__________________________________________________________________________________________________
block_2a_conv_shortcut (Conv2D) (None, 128, 48, 156) 8192 block_1b_relu[0][0]
__________________________________________________________________________________________________
block_2a_bn_2 (BatchNormalizati (None, 128, 48, 156) 512 block_2a_conv_2[0][0]
__________________________________________________________________________________________________
block_2a_bn_shortcut (BatchNorm (None, 128, 48, 156) 512 block_2a_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_3 (Add) (None, 128, 48, 156) 0 block_2a_bn_2[0][0]
block_2a_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_2a_relu (Activation) (None, 128, 48, 156) 0 add_3[0][0]
__________________________________________________________________________________________________
block_2b_conv_1 (Conv2D) (None, 128, 48, 156) 147456 block_2a_relu[0][0]
__________________________________________________________________________________________________
block_2b_bn_1 (BatchNormalizati (None, 128, 48, 156) 512 block_2b_conv_1[0][0]
__________________________________________________________________________________________________
block_2b_relu_1 (Activation) (None, 128, 48, 156) 0 block_2b_bn_1[0][0]
__________________________________________________________________________________________________
block_2b_conv_2 (Conv2D) (None, 128, 48, 156) 147456 block_2b_relu_1[0][0]
__________________________________________________________________________________________________
block_2b_conv_shortcut (Conv2D) (None, 128, 48, 156) 16384 block_2a_relu[0][0]
__________________________________________________________________________________________________
block_2b_bn_2 (BatchNormalizati (None, 128, 48, 156) 512 block_2b_conv_2[0][0]
__________________________________________________________________________________________________
block_2b_bn_shortcut (BatchNorm (None, 128, 48, 156) 512 block_2b_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_4 (Add) (None, 128, 48, 156) 0 block_2b_bn_2[0][0]
block_2b_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_2b_relu (Activation) (None, 128, 48, 156) 0 add_4[0][0]
__________________________________________________________________________________________________
block_3a_conv_1 (Conv2D) (None, 256, 24, 78) 294912 block_2b_relu[0][0]
__________________________________________________________________________________________________
block_3a_bn_1 (BatchNormalizati (None, 256, 24, 78) 1024 block_3a_conv_1[0][0]
__________________________________________________________________________________________________
block_3a_relu_1 (Activation) (None, 256, 24, 78) 0 block_3a_bn_1[0][0]
__________________________________________________________________________________________________
block_3a_conv_2 (Conv2D) (None, 256, 24, 78) 589824 block_3a_relu_1[0][0]
__________________________________________________________________________________________________
block_3a_conv_shortcut (Conv2D) (None, 256, 24, 78) 32768 block_2b_relu[0][0]
__________________________________________________________________________________________________
block_3a_bn_2 (BatchNormalizati (None, 256, 24, 78) 1024 block_3a_conv_2[0][0]
__________________________________________________________________________________________________
block_3a_bn_shortcut (BatchNorm (None, 256, 24, 78) 1024 block_3a_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_5 (Add) (None, 256, 24, 78) 0 block_3a_bn_2[0][0]
block_3a_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_3a_relu (Activation) (None, 256, 24, 78) 0 add_5[0][0]
__________________________________________________________________________________________________
block_3b_conv_1 (Conv2D) (None, 256, 24, 78) 589824 block_3a_relu[0][0]
__________________________________________________________________________________________________
block_3b_bn_1 (BatchNormalizati (None, 256, 24, 78) 1024 block_3b_conv_1[0][0]
__________________________________________________________________________________________________
block_3b_relu_1 (Activation) (None, 256, 24, 78) 0 block_3b_bn_1[0][0]
__________________________________________________________________________________________________
block_3b_conv_2 (Conv2D) (None, 256, 24, 78) 589824 block_3b_relu_1[0][0]
__________________________________________________________________________________________________
block_3b_conv_shortcut (Conv2D) (None, 256, 24, 78) 65536 block_3a_relu[0][0]
__________________________________________________________________________________________________
block_3b_bn_2 (BatchNormalizati (None, 256, 24, 78) 1024 block_3b_conv_2[0][0]
__________________________________________________________________________________________________
block_3b_bn_shortcut (BatchNorm (None, 256, 24, 78) 1024 block_3b_conv_shortcut[0][0]
__________________________________________________________________________________________________
add_6 (Add) (None, 256, 24, 78) 0 block_3b_bn_2[0][0]
block_3b_bn_shortcut[0][0]
__________________________________________________________________________________________________
block_3b_relu (Activation) (None, 256, 24, 78) 0 add_6[0][0]
__________________________________________________________________________________________________
rpn_conv1 (Conv2D) (None, 512, 24, 78) 1180160 block_3b_relu[0][0]
__________________________________________________________________________________________________
rpn_out_class (Conv2D) (None, 9, 24, 78) 4617 rpn_conv1[0][0]
__________________________________________________________________________________________________
rpn_out_regress (Conv2D) (None, 36, 24, 78) 18468 rpn_conv1[0][0]
__________________________________________________________________________________________________
proposal_1 (Proposal) (None, 300, 4) 0 rpn_out_class[0][0]
rpn_out_regress[0][0]
input_image[0][0]
__________________________________________________________________________________________________
crop_and_resize_1 (CropAndResiz (None, 300, 256, 7, 0 block_3b_relu[0][0]
proposal_1[0][0]
input_image[0][0]
__________________________________________________________________________________________________
time_distributed_1 (TimeDistrib (None, 300, 512, 7, 1179648 crop_and_resize_1[0][0]
__________________________________________________________________________________________________
time_distributed_2 (TimeDistrib (None, 300, 512, 7, 2048 time_distributed_1[0][0]
__________________________________________________________________________________________________
block_4a_relu_1 (Activation) (None, 300, 512, 7, 0 time_distributed_2[0][0]
__________________________________________________________________________________________________
time_distributed_3 (TimeDistrib (None, 300, 512, 7, 2359296 block_4a_relu_1[0][0]
__________________________________________________________________________________________________
time_distributed_5 (TimeDistrib (None, 300, 512, 7, 131072 crop_and_resize_1[0][0]
__________________________________________________________________________________________________
time_distributed_4 (TimeDistrib (None, 300, 512, 7, 2048 time_distributed_3[0][0]
__________________________________________________________________________________________________
time_distributed_6 (TimeDistrib (None, 300, 512, 7, 2048 time_distributed_5[0][0]
__________________________________________________________________________________________________
add_7 (Add) (None, 300, 512, 7, 0 time_distributed_4[0][0]
time_distributed_6[0][0]
__________________________________________________________________________________________________
block_4a_relu (Activation) (None, 300, 512, 7, 0 add_7[0][0]
__________________________________________________________________________________________________
time_distributed_7 (TimeDistrib (None, 300, 512, 7, 2359296 block_4a_relu[0][0]
__________________________________________________________________________________________________
time_distributed_8 (TimeDistrib (None, 300, 512, 7, 2048 time_distributed_7[0][0]
__________________________________________________________________________________________________
block_4b_relu_1 (Activation) (None, 300, 512, 7, 0 time_distributed_8[0][0]
__________________________________________________________________________________________________
time_distributed_9 (TimeDistrib (None, 300, 512, 7, 2359296 block_4b_relu_1[0][0]
__________________________________________________________________________________________________
time_distributed_11 (TimeDistri (None, 300, 512, 7, 262144 block_4a_relu[0][0]
__________________________________________________________________________________________________
time_distributed_10 (TimeDistri (None, 300, 512, 7, 2048 time_distributed_9[0][0]
__________________________________________________________________________________________________
time_distributed_12 (TimeDistri (None, 300, 512, 7, 2048 time_distributed_11[0][0]
__________________________________________________________________________________________________
add_8 (Add) (None, 300, 512, 7, 0 time_distributed_10[0][0]
time_distributed_12[0][0]
__________________________________________________________________________________________________
block_4b_relu (Activation) (None, 300, 512, 7, 0 add_8[0][0]
__________________________________________________________________________________________________
time_distributed_13 (TimeDistri (None, 300, 512, 1, 0 block_4b_relu[0][0]
__________________________________________________________________________________________________
time_distributed_flatten (TimeD (None, 300, 512) 0 time_distributed_13[0][0]
__________________________________________________________________________________________________
dense_class_td (TimeDistributed (None, 300, 4) 2052 time_distributed_flatten[0][0]
__________________________________________________________________________________________________
dense_regress_td (TimeDistribut (None, 300, 12) 6156 time_distributed_flatten[0][0]
__________________________________________________________________________________________________
nms_inputs_1 (NmsInputs) [(None, 4800, 1, 1), 0 proposal_1[0][0]
dense_class_td[0][0]
dense_regress_td[0][0]
==================================================================================================
Total params: 12,741,261
Trainable params: 12,577,181
Non-trainable params: 164,080
__________________________________________________________________________________________________
NOTE: UFF has been tested with TensorFlow 1.14.0.
WARNING: The version of TensorFlow installed on this system is not guaranteed to work with UFF.
Warning: No conversion function registered for layer: NMS_TRT yet.
Converting NMS as custom op: NMS_TRT
Warning: No conversion function registered for layer: Proposal yet.
Converting proposal as custom op: Proposal
DEBUG: convert reshape to flatten node
Warning: No conversion function registered for layer: CropAndResize yet.
Converting roi_pooling_conv_1/CropAndResize_new as custom op: CropAndResize
DEBUG [/usr/local/lib/python3.6/dist-packages/uff/converters/tensorflow/converter.py:96] Marking ['NMS'] as outputs
Traceback (most recent call last):
File "/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/faster_rcnn/scripts/export.py", line 12, in <module>
File "/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/common/export/app.py", line 219, in launch_export
File "/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/common/export/app.py", line 201, in run_export
File "/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/common/export/keras_exporter.py", line 360, in export
TypeError: get_calibrator() got an unexpected keyword argument 'image_mean'
root@f3250a75bdd4:/workspace#