I installed Tensorflow 1.7 with TensorRT integration from https://devtalk.nvidia.com/default/topic/1031300/jetson-tx2/tensorflow-1-7-wheel-with-jetpack-3-2-/2#reply. I tried running a simple code, but the it fails with
2018-04-05 11:48:57.688753: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:865] ARM64 does not support NUMA - returning NUMA node zero 2018-04-05 11:48:57.688893: I tensorflow/core/grappler/devices.cc:51] Number of eligible GPUs (core count >= 8): 0 terminate called after throwing an instance of 'std::out_of_range' what(): basic_string::substr: __pos (which is 13) > this->size() (which is 12) Aborted (core dumped)
For better explanation the code that should load the model.
self.__graph_def = self.load_graphdef(graphfilename) self.__graph = trt.create_inference_graph(self.__graph_def, outputs=["Argmax_Image:0"], max_batch_size=1, max_workspace_size_bytes=2000000000, precision_mode="FP16") self.__x = self.__graph.get_tensor_by_name("Input:0") self.__y = self.__graph.get_tensor_by_name("Argmax_Image:0") gpu_options = tf.GPUOptions(per_process_gpu_memory_fraction=0.90) self.__sess = tf.Session(graph=self.__graph, config=tf.ConfigProto(gpu_options=gpu_options))
Has anyone had an similar issue or an idea how to resolve it?