We have a task of improving model inferencing time. Ofcourse on of the solution could be gpu distribution training but we don’t have feasibility to go with this approach for various reasons. So we were thinking of improving inferencing time by optimising low level design by using cython for hardware acceleration on jetson device. So just wanted to know which area I should Target? If any url to study or any idea would really be appreciated.
If this seems very generic, then please ask me the specific topic so that I can provide more information which inturn will help to get correct approach.
I took the liberty now to get you started with the Jetson community, since any inference optimizations you would want to try out should start with the chosen HW environment.
But I suggest you also have a look in the Frameworks category, maybe look for content tagged with pytorch or jetson-inference.