2022-07-29 05:10:16,104 [INFO] root: Registry: ['nvcr.io'] 2022-07-29 05:10:16,249 [INFO] tlt.components.instance_handler.local_instance: Running command in container: nvcr.io/nvidia/tao/tao-toolkit-tf:v3.21.11-tf1.15.5-py3 Matplotlib created a temporary config/cache directory at /tmp/matplotlib-xpr8bc1e because the default path (/.config/matplotlib) is not a writable directory; it is highly recommended to set the MPLCONFIGDIR environment variable to a writable directory, in particular to speed up the import of Matplotlib and to better support multiprocessing. Using TensorFlow backend. WARNING:tensorflow:Deprecation warnings have been disabled. Set TF_ENABLE_DEPRECATION_WARNINGS=1 to re-enable them. Using TensorFlow backend. 2022-07-29 09:10:23,510 [INFO] iva.unet.spec_handler.spec_loader: Merging specification from /workspace/tao-experiments/specs/unet_train_vgg_6S.txt WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:153: The name tf.get_default_graph is deprecated. Please use tf.compat.v1.get_default_graph instead. 2022-07-29 09:10:23,513 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:153: The name tf.get_default_graph is deprecated. Please use tf.compat.v1.get_default_graph instead. 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 0: Train Id 0 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 1: Train Id 1 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 2: Train Id 2 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 3: Train Id 3 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 4: Train Id 4 2022-07-29 09:10:23,514 [INFO] iva.unet.model.utilities: Label Id 5: Train Id 5 2022-07-29 09:10:23,515 [INFO] iva.unet.model.model_io: Loading weights from /workspace/tao-experiments/unpruned/weights/model.tlt WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:517: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead. 2022-07-29 09:10:27,487 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:517: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:4138: The name tf.random_uniform is deprecated. Please use tf.random.uniform instead. 2022-07-29 09:10:27,489 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:4138: The name tf.random_uniform is deprecated. Please use tf.random.uniform instead. WARNING:tensorflow:From /opt/nvidia/third_party/keras/tensorflow_backend.py:183: The name tf.nn.max_pool is deprecated. Please use tf.nn.max_pool2d instead. 2022-07-29 09:10:27,518 [WARNING] tensorflow: From /opt/nvidia/third_party/keras/tensorflow_backend.py:183: The name tf.nn.max_pool is deprecated. Please use tf.nn.max_pool2d instead. __________________________________________________________________________________________________ Layer (type) Output Shape Param # Connected to ================================================================================================== input_1 (InputLayer) (None, 3, 704, 1280) 0 __________________________________________________________________________________________________ block_1a_conv_1 (Conv2D) (None, 64, 704, 1280 1792 input_1[0][0] __________________________________________________________________________________________________ block_1a_relu (Activation) (None, 64, 704, 1280 0 block_1a_conv_1[0][0] __________________________________________________________________________________________________ block_1b_conv_1 (Conv2D) (None, 64, 704, 1280 36928 block_1a_relu[0][0] __________________________________________________________________________________________________ block_1b_relu (Activation) (None, 64, 704, 1280 0 block_1b_conv_1[0][0] __________________________________________________________________________________________________ block1_pool (MaxPooling2D) (None, 64, 352, 640) 0 block_1b_relu[0][0] __________________________________________________________________________________________________ block_2a_conv_1 (Conv2D) (None, 128, 352, 640 73856 block1_pool[0][0] __________________________________________________________________________________________________ block_2a_relu (Activation) (None, 128, 352, 640 0 block_2a_conv_1[0][0] __________________________________________________________________________________________________ block_2b_conv_1 (Conv2D) (None, 128, 352, 640 147584 block_2a_relu[0][0] __________________________________________________________________________________________________ block_2b_relu (Activation) (None, 128, 352, 640 0 block_2b_conv_1[0][0] __________________________________________________________________________________________________ block2_pool (MaxPooling2D) (None, 128, 176, 320 0 block_2b_relu[0][0] __________________________________________________________________________________________________ block_3a_conv_1 (Conv2D) (None, 256, 176, 320 295168 block2_pool[0][0] __________________________________________________________________________________________________ block_3a_relu (Activation) (None, 256, 176, 320 0 block_3a_conv_1[0][0] __________________________________________________________________________________________________ block_3b_conv_1 (Conv2D) (None, 256, 176, 320 590080 block_3a_relu[0][0] __________________________________________________________________________________________________ block_3b_relu (Activation) (None, 256, 176, 320 0 block_3b_conv_1[0][0] __________________________________________________________________________________________________ block_3c_conv_1 (Conv2D) (None, 256, 176, 320 590080 block_3b_relu[0][0] __________________________________________________________________________________________________ block_3c_relu (Activation) (None, 256, 176, 320 0 block_3c_conv_1[0][0] __________________________________________________________________________________________________ block3_pool (MaxPooling2D) (None, 256, 88, 160) 0 block_3c_relu[0][0] __________________________________________________________________________________________________ block_4a_conv_1 (Conv2D) (None, 512, 88, 160) 1180160 block3_pool[0][0] __________________________________________________________________________________________________ block_4a_relu (Activation) (None, 512, 88, 160) 0 block_4a_conv_1[0][0] __________________________________________________________________________________________________ block_4b_conv_1 (Conv2D) (None, 512, 88, 160) 2359808 block_4a_relu[0][0] __________________________________________________________________________________________________ block_4b_relu (Activation) (None, 512, 88, 160) 0 block_4b_conv_1[0][0] __________________________________________________________________________________________________ block_4c_conv_1 (Conv2D) (None, 512, 88, 160) 2359808 block_4b_relu[0][0] __________________________________________________________________________________________________ block_4c_relu (Activation) (None, 512, 88, 160) 0 block_4c_conv_1[0][0] __________________________________________________________________________________________________ block4_pool (MaxPooling2D) (None, 512, 44, 80) 0 block_4c_relu[0][0] __________________________________________________________________________________________________ block_5a_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block4_pool[0][0] __________________________________________________________________________________________________ block_5a_relu (Activation) (None, 512, 44, 80) 0 block_5a_conv_1[0][0] __________________________________________________________________________________________________ block_5b_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block_5a_relu[0][0] __________________________________________________________________________________________________ block_5b_relu (Activation) (None, 512, 44, 80) 0 block_5b_conv_1[0][0] __________________________________________________________________________________________________ block_5c_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block_5b_relu[0][0] __________________________________________________________________________________________________ block_5c_relu (Activation) (None, 512, 44, 80) 0 block_5c_conv_1[0][0] __________________________________________________________________________________________________ max_pooling2d_1 (MaxPooling2D) (None, 512, 22, 40) 0 block_5c_relu[0][0] __________________________________________________________________________________________________ conv2d_transpose_1 (Conv2DTrans (None, 512, 44, 80) 4194816 max_pooling2d_1[0][0] __________________________________________________________________________________________________ concatenate_1 (Concatenate) (None, 1024, 44, 80) 0 conv2d_transpose_1[0][0] block4_pool[0][0] __________________________________________________________________________________________________ conv2d_1 (Conv2D) (None, 512, 44, 80) 4719104 concatenate_1[0][0] __________________________________________________________________________________________________ conv2d_2 (Conv2D) (None, 512, 44, 80) 2359808 conv2d_1[0][0] __________________________________________________________________________________________________ conv2d_transpose_2 (Conv2DTrans (None, 256, 88, 160) 2097408 conv2d_2[0][0] __________________________________________________________________________________________________ concatenate_2 (Concatenate) (None, 512, 88, 160) 0 conv2d_transpose_2[0][0] block3_pool[0][0] __________________________________________________________________________________________________ conv2d_3 (Conv2D) (None, 256, 88, 160) 1179904 concatenate_2[0][0] __________________________________________________________________________________________________ conv2d_4 (Conv2D) (None, 256, 88, 160) 590080 conv2d_3[0][0] __________________________________________________________________________________________________ conv2d_transpose_3 (Conv2DTrans (None, 128, 176, 320 524416 conv2d_4[0][0] __________________________________________________________________________________________________ concatenate_3 (Concatenate) (None, 256, 176, 320 0 conv2d_transpose_3[0][0] block2_pool[0][0] __________________________________________________________________________________________________ conv2d_5 (Conv2D) (None, 128, 176, 320 295040 concatenate_3[0][0] __________________________________________________________________________________________________ conv2d_6 (Conv2D) (None, 128, 176, 320 147584 conv2d_5[0][0] __________________________________________________________________________________________________ conv2d_transpose_4 (Conv2DTrans (None, 64, 352, 640) 131136 conv2d_6[0][0] __________________________________________________________________________________________________ concatenate_4 (Concatenate) (None, 128, 352, 640 0 conv2d_transpose_4[0][0] block1_pool[0][0] __________________________________________________________________________________________________ conv2d_7 (Conv2D) (None, 64, 352, 640) 73792 concatenate_4[0][0] __________________________________________________________________________________________________ conv2d_8 (Conv2D) (None, 64, 352, 640) 36928 conv2d_7[0][0] __________________________________________________________________________________________________ conv2d_transpose_5 (Conv2DTrans (None, 64, 704, 1280 65600 conv2d_8[0][0] __________________________________________________________________________________________________ concatenate_5 (Concatenate) (None, 128, 704, 128 0 conv2d_transpose_5[0][0] block_1a_relu[0][0] __________________________________________________________________________________________________ conv2d_9 (Conv2D) (None, 64, 704, 1280 73792 concatenate_5[0][0] __________________________________________________________________________________________________ conv2d_10 (Conv2D) (None, 64, 704, 1280 36928 conv2d_9[0][0] __________________________________________________________________________________________________ conv2d_11 (Conv2D) (None, 6, 704, 1280) 390 conv2d_10[0][0] ================================================================================================== Total params: 31,241,414 Trainable params: 31,241,414 Non-trainable params: 0 __________________________________________________________________________________________________ WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:174: The name tf.get_default_session is deprecated. Please use tf.compat.v1.get_default_session instead. 2022-07-29 09:10:27,828 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:174: The name tf.get_default_session is deprecated. Please use tf.compat.v1.get_default_session instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:181: The name tf.ConfigProto is deprecated. Please use tf.compat.v1.ConfigProto instead. 2022-07-29 09:10:27,828 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:181: The name tf.ConfigProto is deprecated. Please use tf.compat.v1.ConfigProto instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:186: The name tf.Session is deprecated. Please use tf.compat.v1.Session instead. 2022-07-29 09:10:27,828 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:186: The name tf.Session is deprecated. Please use tf.compat.v1.Session instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:190: The name tf.global_variables is deprecated. Please use tf.compat.v1.global_variables instead. 2022-07-29 09:10:28,233 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:190: The name tf.global_variables is deprecated. Please use tf.compat.v1.global_variables instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:199: The name tf.is_variable_initialized is deprecated. Please use tf.compat.v1.is_variable_initialized instead. 2022-07-29 09:10:28,233 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:199: The name tf.is_variable_initialized is deprecated. Please use tf.compat.v1.is_variable_initialized instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:206: The name tf.variables_initializer is deprecated. Please use tf.compat.v1.variables_initializer instead. 2022-07-29 09:10:28,577 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:206: The name tf.variables_initializer is deprecated. Please use tf.compat.v1.variables_initializer instead. WARNING:tensorflow:From /root/.cache/bazel/_bazel_root/ed34e6d125608f91724fda23656f1726/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/unet/model/model_io.py:71: The name tf.global_variables_initializer is deprecated. Please use tf.compat.v1.global_variables_initializer instead. 2022-07-29 09:10:28,695 [WARNING] tensorflow: From /root/.cache/bazel/_bazel_root/ed34e6d125608f91724fda23656f1726/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/iva/build_wheel.runfiles/ai_infra/iva/unet/model/model_io.py:71: The name tf.global_variables_initializer is deprecated. Please use tf.compat.v1.global_variables_initializer instead. INFO:tensorflow:Restoring parameters from /tmp/tmpnjqvhsif/model.ckpt-20000 2022-07-29 09:10:28,735 [INFO] tensorflow: Restoring parameters from /tmp/tmpnjqvhsif/model.ckpt-20000 2022-07-29 09:10:28,928 [INFO] iva.unet.model.model_io: Loaded weights Successfully for Export 2022-07-29 09:10:28,928 [INFO] __main__: Original unpruned Model summary: __________________________________________________________________________________________________ Layer (type) Output Shape Param # Connected to ================================================================================================== input_1 (InputLayer) (None, 3, 704, 1280) 0 __________________________________________________________________________________________________ block_1a_conv_1 (Conv2D) (None, 64, 704, 1280 1792 input_1[0][0] __________________________________________________________________________________________________ block_1a_relu (Activation) (None, 64, 704, 1280 0 block_1a_conv_1[0][0] __________________________________________________________________________________________________ block_1b_conv_1 (Conv2D) (None, 64, 704, 1280 36928 block_1a_relu[0][0] __________________________________________________________________________________________________ block_1b_relu (Activation) (None, 64, 704, 1280 0 block_1b_conv_1[0][0] __________________________________________________________________________________________________ block1_pool (MaxPooling2D) (None, 64, 352, 640) 0 block_1b_relu[0][0] __________________________________________________________________________________________________ block_2a_conv_1 (Conv2D) (None, 128, 352, 640 73856 block1_pool[0][0] __________________________________________________________________________________________________ block_2a_relu (Activation) (None, 128, 352, 640 0 block_2a_conv_1[0][0] __________________________________________________________________________________________________ block_2b_conv_1 (Conv2D) (None, 128, 352, 640 147584 block_2a_relu[0][0] __________________________________________________________________________________________________ block_2b_relu (Activation) (None, 128, 352, 640 0 block_2b_conv_1[0][0] __________________________________________________________________________________________________ block2_pool (MaxPooling2D) (None, 128, 176, 320 0 block_2b_relu[0][0] __________________________________________________________________________________________________ block_3a_conv_1 (Conv2D) (None, 256, 176, 320 295168 block2_pool[0][0] __________________________________________________________________________________________________ block_3a_relu (Activation) (None, 256, 176, 320 0 block_3a_conv_1[0][0] __________________________________________________________________________________________________ block_3b_conv_1 (Conv2D) (None, 256, 176, 320 590080 block_3a_relu[0][0] __________________________________________________________________________________________________ block_3b_relu (Activation) (None, 256, 176, 320 0 block_3b_conv_1[0][0] __________________________________________________________________________________________________ block_3c_conv_1 (Conv2D) (None, 256, 176, 320 590080 block_3b_relu[0][0] __________________________________________________________________________________________________ block_3c_relu (Activation) (None, 256, 176, 320 0 block_3c_conv_1[0][0] __________________________________________________________________________________________________ block3_pool (MaxPooling2D) (None, 256, 88, 160) 0 block_3c_relu[0][0] __________________________________________________________________________________________________ block_4a_conv_1 (Conv2D) (None, 512, 88, 160) 1180160 block3_pool[0][0] __________________________________________________________________________________________________ block_4a_relu (Activation) (None, 512, 88, 160) 0 block_4a_conv_1[0][0] __________________________________________________________________________________________________ block_4b_conv_1 (Conv2D) (None, 512, 88, 160) 2359808 block_4a_relu[0][0] __________________________________________________________________________________________________ block_4b_relu (Activation) (None, 512, 88, 160) 0 block_4b_conv_1[0][0] __________________________________________________________________________________________________ block_4c_conv_1 (Conv2D) (None, 512, 88, 160) 2359808 block_4b_relu[0][0] __________________________________________________________________________________________________ block_4c_relu (Activation) (None, 512, 88, 160) 0 block_4c_conv_1[0][0] __________________________________________________________________________________________________ block4_pool (MaxPooling2D) (None, 512, 44, 80) 0 block_4c_relu[0][0] __________________________________________________________________________________________________ block_5a_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block4_pool[0][0] __________________________________________________________________________________________________ block_5a_relu (Activation) (None, 512, 44, 80) 0 block_5a_conv_1[0][0] __________________________________________________________________________________________________ block_5b_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block_5a_relu[0][0] __________________________________________________________________________________________________ block_5b_relu (Activation) (None, 512, 44, 80) 0 block_5b_conv_1[0][0] __________________________________________________________________________________________________ block_5c_conv_1 (Conv2D) (None, 512, 44, 80) 2359808 block_5b_relu[0][0] __________________________________________________________________________________________________ block_5c_relu (Activation) (None, 512, 44, 80) 0 block_5c_conv_1[0][0] __________________________________________________________________________________________________ max_pooling2d_1 (MaxPooling2D) (None, 512, 22, 40) 0 block_5c_relu[0][0] __________________________________________________________________________________________________ conv2d_transpose_1 (Conv2DTrans (None, 512, 44, 80) 4194816 max_pooling2d_1[0][0] __________________________________________________________________________________________________ concatenate_1 (Concatenate) (None, 1024, 44, 80) 0 conv2d_transpose_1[0][0] block4_pool[0][0] __________________________________________________________________________________________________ conv2d_1 (Conv2D) (None, 512, 44, 80) 4719104 concatenate_1[0][0] __________________________________________________________________________________________________ conv2d_2 (Conv2D) (None, 512, 44, 80) 2359808 conv2d_1[0][0] __________________________________________________________________________________________________ conv2d_transpose_2 (Conv2DTrans (None, 256, 88, 160) 2097408 conv2d_2[0][0] __________________________________________________________________________________________________ concatenate_2 (Concatenate) (None, 512, 88, 160) 0 conv2d_transpose_2[0][0] block3_pool[0][0] __________________________________________________________________________________________________ conv2d_3 (Conv2D) (None, 256, 88, 160) 1179904 concatenate_2[0][0] __________________________________________________________________________________________________ conv2d_4 (Conv2D) (None, 256, 88, 160) 590080 conv2d_3[0][0] __________________________________________________________________________________________________ conv2d_transpose_3 (Conv2DTrans (None, 128, 176, 320 524416 conv2d_4[0][0] __________________________________________________________________________________________________ concatenate_3 (Concatenate) (None, 256, 176, 320 0 conv2d_transpose_3[0][0] block2_pool[0][0] __________________________________________________________________________________________________ conv2d_5 (Conv2D) (None, 128, 176, 320 295040 concatenate_3[0][0] __________________________________________________________________________________________________ conv2d_6 (Conv2D) (None, 128, 176, 320 147584 conv2d_5[0][0] __________________________________________________________________________________________________ conv2d_transpose_4 (Conv2DTrans (None, 64, 352, 640) 131136 conv2d_6[0][0] __________________________________________________________________________________________________ concatenate_4 (Concatenate) (None, 128, 352, 640 0 conv2d_transpose_4[0][0] block1_pool[0][0] __________________________________________________________________________________________________ conv2d_7 (Conv2D) (None, 64, 352, 640) 73792 concatenate_4[0][0] __________________________________________________________________________________________________ conv2d_8 (Conv2D) (None, 64, 352, 640) 36928 conv2d_7[0][0] __________________________________________________________________________________________________ conv2d_transpose_5 (Conv2DTrans (None, 64, 704, 1280 65600 conv2d_8[0][0] __________________________________________________________________________________________________ concatenate_5 (Concatenate) (None, 128, 704, 128 0 conv2d_transpose_5[0][0] block_1a_relu[0][0] __________________________________________________________________________________________________ conv2d_9 (Conv2D) (None, 64, 704, 1280 73792 concatenate_5[0][0] __________________________________________________________________________________________________ conv2d_10 (Conv2D) (None, 64, 704, 1280 36928 conv2d_9[0][0] __________________________________________________________________________________________________ conv2d_11 (Conv2D) (None, 6, 704, 1280) 390 conv2d_10[0][0] ================================================================================================== Total params: 31,241,414 Trainable params: 31,241,414 Non-trainable params: 0 __________________________________________________________________________________________________ 2022-07-29 09:10:28,950 [INFO] modulus.pruning.pruning: Exploring graph for retainable indices 2022-07-29 09:10:29,385 [INFO] modulus.pruning.pruning: Pruning model and appending pruned nodes to new graph 2022-07-29 09:10:34,403 [INFO] __main__: Model summary of the pruned model: __________________________________________________________________________________________________ Layer (type) Output Shape Param # Connected to ================================================================================================== input_1 (InputLayer) (None, 3, 704, 1280) 0 __________________________________________________________________________________________________ block_1a_conv_1 (Conv2D) (None, 64, 704, 1280 1792 input_1[0][0] __________________________________________________________________________________________________ block_1a_relu (Activation) (None, 64, 704, 1280 0 block_1a_conv_1[0][0] __________________________________________________________________________________________________ block_1b_conv_1 (Conv2D) (None, 64, 704, 1280 36928 block_1a_relu[0][0] __________________________________________________________________________________________________ block_1b_relu (Activation) (None, 64, 704, 1280 0 block_1b_conv_1[0][0] __________________________________________________________________________________________________ block1_pool (MaxPooling2D) (None, 64, 352, 640) 0 block_1b_relu[0][0] __________________________________________________________________________________________________ block_2a_conv_1 (Conv2D) (None, 128, 352, 640 73856 block1_pool[0][0] __________________________________________________________________________________________________ block_2a_relu (Activation) (None, 128, 352, 640 0 block_2a_conv_1[0][0] __________________________________________________________________________________________________ block_2b_conv_1 (Conv2D) (None, 112, 352, 640 129136 block_2a_relu[0][0] __________________________________________________________________________________________________ block_2b_relu (Activation) (None, 112, 352, 640 0 block_2b_conv_1[0][0] __________________________________________________________________________________________________ block2_pool (MaxPooling2D) (None, 112, 176, 320 0 block_2b_relu[0][0] __________________________________________________________________________________________________ block_3a_conv_1 (Conv2D) (None, 112, 176, 320 113008 block2_pool[0][0] __________________________________________________________________________________________________ block_3a_relu (Activation) (None, 112, 176, 320 0 block_3a_conv_1[0][0] __________________________________________________________________________________________________ block_3b_conv_1 (Conv2D) (None, 88, 176, 320) 88792 block_3a_relu[0][0] __________________________________________________________________________________________________ block_3b_relu (Activation) (None, 88, 176, 320) 0 block_3b_conv_1[0][0] __________________________________________________________________________________________________ block_3c_conv_1 (Conv2D) (None, 104, 176, 320 82472 block_3b_relu[0][0] __________________________________________________________________________________________________ block_3c_relu (Activation) (None, 104, 176, 320 0 block_3c_conv_1[0][0] __________________________________________________________________________________________________ block3_pool (MaxPooling2D) (None, 104, 88, 160) 0 block_3c_relu[0][0] __________________________________________________________________________________________________ block_4a_conv_1 (Conv2D) (None, 136, 88, 160) 127432 block3_pool[0][0] __________________________________________________________________________________________________ block_4a_relu (Activation) (None, 136, 88, 160) 0 block_4a_conv_1[0][0] __________________________________________________________________________________________________ block_4b_conv_1 (Conv2D) (None, 208, 88, 160) 254800 block_4a_relu[0][0] __________________________________________________________________________________________________ block_4b_relu (Activation) (None, 208, 88, 160) 0 block_4b_conv_1[0][0] __________________________________________________________________________________________________ block_4c_conv_1 (Conv2D) (None, 352, 88, 160) 659296 block_4b_relu[0][0] __________________________________________________________________________________________________ block_4c_relu (Activation) (None, 352, 88, 160) 0 block_4c_conv_1[0][0] __________________________________________________________________________________________________ block4_pool (MaxPooling2D) (None, 352, 44, 80) 0 block_4c_relu[0][0] __________________________________________________________________________________________________ block_5a_conv_1 (Conv2D) (None, 304, 44, 80) 963376 block4_pool[0][0] __________________________________________________________________________________________________ block_5a_relu (Activation) (None, 304, 44, 80) 0 block_5a_conv_1[0][0] __________________________________________________________________________________________________ block_5b_conv_1 (Conv2D) (None, 472, 44, 80) 1291864 block_5a_relu[0][0] __________________________________________________________________________________________________ block_5b_relu (Activation) (None, 472, 44, 80) 0 block_5b_conv_1[0][0] __________________________________________________________________________________________________ block_5c_conv_1 (Conv2D) (None, 408, 44, 80) 1733592 block_5b_relu[0][0] __________________________________________________________________________________________________ block_5c_relu (Activation) (None, 408, 44, 80) 0 block_5c_conv_1[0][0] __________________________________________________________________________________________________ max_pooling2d_1 (MaxPooling2D) (None, 408, 22, 40) 0 block_5c_relu[0][0] __________________________________________________________________________________________________ conv2d_transpose_1 (Conv2DTrans (None, 512, 44, 80) 3342848 max_pooling2d_1[0][0] __________________________________________________________________________________________________ concatenate_1 (Concatenate) (None, 864, 44, 80) 0 conv2d_transpose_1[0][0] block4_pool[0][0] __________________________________________________________________________________________________ conv2d_1 (Conv2D) (None, 112, 44, 80) 871024 concatenate_1[0][0] __________________________________________________________________________________________________ conv2d_2 (Conv2D) (None, 120, 44, 80) 121080 conv2d_1[0][0] __________________________________________________________________________________________________ conv2d_transpose_2 (Conv2DTrans (None, 256, 88, 160) 491776 conv2d_2[0][0] __________________________________________________________________________________________________ concatenate_2 (Concatenate) (None, 360, 88, 160) 0 conv2d_transpose_2[0][0] block3_pool[0][0] __________________________________________________________________________________________________ conv2d_3 (Conv2D) (None, 80, 88, 160) 259280 concatenate_2[0][0] __________________________________________________________________________________________________ conv2d_4 (Conv2D) (None, 56, 88, 160) 40376 conv2d_3[0][0] __________________________________________________________________________________________________ conv2d_transpose_3 (Conv2DTrans (None, 128, 176, 320 114816 conv2d_4[0][0] __________________________________________________________________________________________________ concatenate_3 (Concatenate) (None, 240, 176, 320 0 conv2d_transpose_3[0][0] block2_pool[0][0] __________________________________________________________________________________________________ conv2d_5 (Conv2D) (None, 104, 176, 320 224744 concatenate_3[0][0] __________________________________________________________________________________________________ conv2d_6 (Conv2D) (None, 128, 176, 320 119936 conv2d_5[0][0] __________________________________________________________________________________________________ conv2d_transpose_4 (Conv2DTrans (None, 64, 352, 640) 131136 conv2d_6[0][0] __________________________________________________________________________________________________ concatenate_4 (Concatenate) (None, 128, 352, 640 0 conv2d_transpose_4[0][0] block1_pool[0][0] __________________________________________________________________________________________________ conv2d_7 (Conv2D) (None, 64, 352, 640) 73792 concatenate_4[0][0] __________________________________________________________________________________________________ conv2d_8 (Conv2D) (None, 64, 352, 640) 36928 conv2d_7[0][0] __________________________________________________________________________________________________ conv2d_transpose_5 (Conv2DTrans (None, 64, 704, 1280 65600 conv2d_8[0][0] __________________________________________________________________________________________________ concatenate_5 (Concatenate) (None, 128, 704, 128 0 conv2d_transpose_5[0][0] block_1a_relu[0][0] __________________________________________________________________________________________________ conv2d_9 (Conv2D) (None, 64, 704, 1280 73792 concatenate_5[0][0] __________________________________________________________________________________________________ conv2d_10 (Conv2D) (None, 64, 704, 1280 36928 conv2d_9[0][0] __________________________________________________________________________________________________ conv2d_11 (Conv2D) (None, 6, 704, 1280) 390 conv2d_10[0][0] ================================================================================================== Total params: 11,560,790 Trainable params: 11,560,790 Non-trainable params: 0 __________________________________________________________________________________________________ 2022-07-29 09:10:34,408 [INFO] __main__: Pruning ratio (pruned model / original model): 0.3700469511399196 WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:95: The name tf.reset_default_graph is deprecated. Please use tf.compat.v1.reset_default_graph instead. 2022-07-29 09:10:34,689 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:95: The name tf.reset_default_graph is deprecated. Please use tf.compat.v1.reset_default_graph instead. WARNING:tensorflow:From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:98: The name tf.placeholder_with_default is deprecated. Please use tf.compat.v1.placeholder_with_default instead. 2022-07-29 09:10:34,715 [WARNING] tensorflow: From /usr/local/lib/python3.6/dist-packages/keras/backend/tensorflow_backend.py:98: The name tf.placeholder_with_default is deprecated. Please use tf.compat.v1.placeholder_with_default instead. 2022-07-29 05:10:38,682 [INFO] tlt.components.docker_handler.docker_handler: Stopping container.