seems like a resource/semaphore race condition. To help us debug, can you share a small repro that contains the model and clients that demonstrate the error you are seeing?
Yes I can. Where I can upload the model and client (python file)?
I am use: docker tensorrtserver:19.01-py3
I build the InferenceClient: from 19.01 branch