For any model I am using, before trying to parse it, first I check if it fully supported using the parser supports_model API and only if its returned status is True I move to the next step which is the activation of parse_from_file API which always return False.
I masked the call to supports_model and surprisingly the problem disappeared and the parse_from_file API started to return True.
When I reactivate supports_model API again before calling to the parse_from_file API the problem immediately return .
No matter if I activate the supports_model API, the parse API returned status is True and the execute_async_v2 service is working well.
When I declare two separated sets of: builder, network and parser, one for supports_model and second for inference execution flow (parse, engine and context creation and finally execution…) everything is working well.
TensorRT Version: 18.104.22.168
GPU Type: Quadro RTX 3000
Nvidia Driver Version: R471.11
CUDA Version: 11.2
CUDNN Version: 8.1.1
Operating System + Version: Windows 10
Python Version (if applicable): 3.6.8
TensorFlow Version (if applicable): NA
PyTorch Version (if applicable): NA
Baremetal or Container (if container which image + tag): Baremetal
Please see above description
My questions are:
Why and how the supports_model API impact on the parse_from_file API?
Why if I replace the parse_from_file API with parse API (not duplicated calls but actually replace between them), their returned statues are not the same? If I call to supports_model API before parse API its returned status is True as oppose to parse_from_file API returned status which is False in this scenario.
My expectation is that both of them shall return the same status when the early conditions are the same……Am I right?