I followed the Gazenet notebook instructions, I think it worked, I can train the model. However I was unable to get good results when running the inference.
I was using the pretrained tlt model and the provided MPIIFaceGaze subset. These are what I got:
If I understand correctly, do you mean skip the training and evaluation steps? I just did that and got an error ‘Please provide a valid gazenet file path for evaluation.’ Or is there anything need to be changed in the spec yaml file? Thank you.
Traceback (most recent call last):
File “/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/driveix/build_wheel.runfiles/ai_infra/driveix/mitgazenet/scripts/inference.py”, line 121, in
File “/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/driveix/build_wheel.runfiles/ai_infra/driveix/mitgazenet/scripts/inference.py”, line 117, in main
File “/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/driveix/build_wheel.runfiles/ai_infra/driveix/mitgazenet/inferencer/gazenet_inferencer.py”, line 97, in build_model
ValueError: Please provide a valid gazenet file path for evaluation.
Traceback (most recent call last):
File “/usr/local/bin/gazenet”, line 8, in
sys.exit(main())
File “/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/driveix/build_wheel.runfiles/ai_infra/driveix/mitgazenet/entrypoint/gazenet.py”, line 13, in main
File “/opt/tlt/.cache/dazel/_dazel_tlt/2b81a5aac84a1d3b7a324f2a7a6f400b/execroot/ai_infra/bazel-out/k8-fastbuild/bin/magnet/packages/driveix/build_wheel.runfiles/ai_infra/driveix/common/entrypoint/entrypoint.py”, line 300, in launch_job
AssertionError: Process run failed.
2021-08-18 11:01:34,342 [INFO] tlt.components.docker_handler.docker_handler: Stopping container.
Yes, skip the training and evaluation steps. Please use the pretrained trainable tlt model and directly run inference with it.
I try on my side, the inference result looks ok.
Please refer to below command in notebook.
Please make sure $LOCAL_EXPERIMENT_DIR/pretrain_models/tlt_gazenet_vtrainable_v1.0/model.tlt is available.
You can run below to verify.
$ tlt unet run ls $LOCAL_EXPERIMENT_DIR/pretrain_models/tlt_gazenet_vtrainable_v1.0/model.tlt
Or you can directly run following command in terminal instead of notebook, to login in the docker.
I run the first command it gives me an error no such file or directory. I thought I followed the notebook to download it, can you tell me how to fix it? Thanks for your patience.