I’m working on a scheduler to allocate image detection inference on either the GPU or CPU. For this, I previously load the model to an object and then start processes for CPU and GPU respectively. When I try to do the inference on the GPU (with cuda), I get a cuda runtime error (801): pytorch\torch/csrc/generic/StorageSharing.cpp:245
I’m not sure why this happens when I decouple the loading from the inference on the GPU. Do you have any idea, how can I solve this?
My setup is:
Python 3.6
Jetson TX1
Pytorch 1.10 - I’m using multiprocessing for a distributed environment.