2018-04-05 11:48:57.688753: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:865] ARM64 does not support NUMA - returning NUMA node zero
2018-04-05 11:48:57.688893: I tensorflow/core/grappler/devices.cc:51] Number of eligible GPUs (core count >= 8): 0
terminate called after throwing an instance of 'std::out_of_range'
what(): basic_string::substr: __pos (which is 13) > this->size() (which is 12)
Aborted (core dumped)
For better explanation the code that should load the model.
trt_graph = trt.create_inference_graph(
input_graph_def=orig_graph,
outputs=["output"],
max_batch_size=inp_dims[0],
max_workspace_size_bytes=1 << 25,
precision_mode="FP32", # TRT Engine precision "FP32","FP16" or "INT8"
minimum_segment_size=2 # minimum number of nodes in an engine
)
(note that 1 << 25 == 33554432, smaller than 2000000000 or 2GB)
Hi kaycc and AastaLLL,
unfortunately the system had a hardware revision and I got sick, hopefully I can test and share the result of the suggestion on Friday.
Hi,
I could now verify the behavious with FP32, but the exact same error as with other precisions.
Yes my model has convolutions with padding=“SAME”. More precisely it is a GoogLeNet FCN model.
We already fixed this issue in TensorRT 4.0 and will be included in our next JetPack release.
If possible, please use padding=‘VALID’ to avoid this bug.
hello AastaLLL, thanks for response. Now i have realize use tensorrt c++ inference to get faster speed by load model in uff format. However, I want to use TensorRT c++ inference to load model from *.pb file just as trt.create_infeerence_graph in python. Do we have correspond function in tensorrt c++ inference?