Hi,
What tools are recommended for doing Model Parallelism for Inference across multiple Jetson devices?
Hi,
What tools are recommended for doing Model Parallelism for Inference across multiple Jetson devices?
For which Jetson platform?
Hi,
Could you share more information about your use case?
Do you want to infer the same input on different platforms for acceleration?
Or do you want to infer multiple inputs and each input run parallel on a device?
Thanks.
For the Orin.
A single input for a model that has weights across devices.
The purpose of using multiple models is for acceleration
Hi,
We don’t have a library to separate the model into difference devices.
This might also be a model-dependent task.
Maybe others can share their experience here.
Thanks.