Hi @rwakker, the tao-toolkit container is for x86. The model training occurs on x86 using TAO, and you can deploy the models trained with TAO to Jetson using DeepStream or Triton Inference Server. For DeepStream on Jetson, you can run the deepstream-l4t container:
The TAO is designed to run on x86 systems with an NVIDIA GPU (e.g., GPU-powered workstation, DGX system, etc) or can be run in any cloud with an NVIDIA GPU. For inference, models can be deployed on any edge device such as an embedded Jetson platform or in a data center with GPUs like T4 or A100, etc.
Giving error:
2022-10-31 20:32:00,786 [INFO] root: Registry: [‘nvcr.io’]
2022-10-31 20:32:00,989 [INFO] tlt.components.instance_handler.local_instance: Running command in container: nvcr.io/nvidia/tao/tao-toolkit-tf:v3.22.05-tf1.15.5-py3
Docker instantiation failed with error: 500 Server Error: Internal Server Error (“failed to create shim: OCI runtime create failed: runc create failed: unable to start container process: error during container init: error running hook #0: error running hook: exit status 1, stdout: , stderr: Auto-detected mode as ‘csv’
invoking the NVIDIA Container Runtime Hook directly (e.g. specifying the docker --gpus flag) is not supported. Please use the NVIDIA Container Runtime instead.: unknown”)
I really like to train my own images, swimmers, where I like to annotate based on keypoints of the partly visible images (swimmers are only seen from top, front view or side view) Arms are lifted up above the water and not visible in the image, etc…
I gave up and now I tried:
This works partly, I need a method to train my data without in-depth knowledge about setting up the complete pipeline,
Configuration of the TAO Toolkit Instance
dockers: [‘nvidia/tao/tao-toolkit-tf’, ‘nvidia/tao/tao-toolkit-pyt’, ‘nvidia/tao/tao-toolkit-lm’]
format_version: 2.0
toolkit_version: 3.22.05
published_date: 05/25/2022
I managed to continue a big step, we can close the topic, it was about the l4t deepstream docker
There is no update from you for a period, assuming this is not an issue anymore. Hence we are closing this topic. If need further support, please open a new one.
Thanks
OK. For training, please run TAO training on x86 systems with dgpu. Cannot run training in Jetson device.
For inference, dgpu or Jetson device is fine.