I want to train a computer vision model for object detection and i’ve seen that NVIDIA already has some Docker containers with models relating to this. Is there any way that I could download all the packages necessary to run this docker container using WSL2? Without having the Jetson/ or connecting it. I would like for the training to be done on a host pc and deploy it to the Jetson Xavier NX. How would I do this?
Hi @dramirez2, yes at one point I had tried PyTorch training under WSL2 using the NGC pytorch container. I used the same train_ssd.py scripts that the jetson-inference Object Detector Training tutorial uses, then copied the exported ONNX model to my Jetson and ran it there through TensorRT with detectnet/detectnet.py.
It may also be possible to run TAO Toolkit for training under WSL2, I’m not sure. Then you could deploy those models on your Jetson with DeepStream.
Presuming that your PC has an NVIDIA GPU, here are some notes I have from setting up WSL2 with docker (these may or may not still be valid)
(reboot if needed)
(wsl --shutdown if needed)
wsl or ubuntu (to start ubuntu)
$ sudo apt-get install docker.io
$ distribution=$(. /etc/os-release;echo $ID$VERSION_ID)
$ curl -s -L https://nvidia.github.io/nvidia-docker/gpgkey | sudo apt-key add -
$ curl -s -L https://nvidia.github.io/nvidia-docker/$distribution/nvidia-docker.list | sudo tee /etc/apt/sources.list.d/nvidia-docker.list
$ sudo apt-get update && sudo apt-get install -y nvidia-container-toolkit
-open a new WSL2 terminal-
# test GPU in container (you probably need to use a newer container tag nowadays)
sudo docker run --rm --gpus all nvidia/cuda:11.0.3-base-ubuntu20.04 nvidia-smi