I’m getting an error when trying to run jarvis_init.sh
. Here are the logs:
Logging into NGC docker registry if necessary...
Pulling required docker images if necessary...
Note: This may take some time, depending on the speed of your Internet connection.
> Pulling Jarvis Speech Server images.
> Image nvcr.io/nvidia/jarvis/jarvis-speech:1.2.1-beta-server exists. Skipping.
> Image nvcr.io/nvidia/jarvis/jarvis-speech-client:1.2.1-beta exists. Skipping.
> Image nvcr.io/nvidia/jarvis/jarvis-speech:1.2.1-beta-servicemaker exists. Skipping.
Converting JMIRs at jarvis-model-repo/jmir to Jarvis Model repository.
+ docker run --shm-size=1g --ulimit memlock=-1 --ulimit stack=67108864 --init -it --rm --gpus '"device=0"' -v jarvis-model-repo:/data -e MODEL_DEPLOY_KEY=tlt_encode --name jarvis-service-maker nvcr.io/nvidia/jarvis/jarvis-speech:1.2.1-beta-servicemaker deploy_all_models /data/jmir /data/models
==========================
== Jarvis Speech Skills ==
==========================
NVIDIA Release devel (build 22382700)
Copyright (c) 2018-2021, NVIDIA CORPORATION. All rights reserved.
Various files include modifications (c) NVIDIA CORPORATION. All rights reserved.
NVIDIA modifications are covered by the license terms that apply to the underlying
project or file.
2021-06-18 01:35:55,409 [INFO] Writing Jarvis model repository to '/data/models'...
2021-06-18 01:35:55,409 [INFO] The jarvis model repo target directory is /data/models
2021-06-18 01:35:56,961 [INFO] Extract_binaries for featurizer -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-feature-extractor-streaming/1
2021-06-18 01:35:56,963 [INFO] Extract_binaries for nn -> /data/models/jarvis-trt-citrinet-1024/1
2021-06-18 01:36:01,092 [INFO] Building TRT engine from ONNX file
[libprotobuf WARNING /workspace/TensorRT/t/oss-cicd/oss/build/third_party.protobuf/src/third_party.protobuf/src/google/protobuf/io/coded_stream.cc:604] Reading dangerously large protocol message. If the message turns out to be larger than 2147483647 bytes, parsing will be halted for security reasons. To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h.
[libprotobuf WARNING /workspace/TensorRT/t/oss-cicd/oss/build/third_party.protobuf/src/third_party.protobuf/src/google/protobuf/io/coded_stream.cc:81] The total number of bytes read was 564429124
[TensorRT] WARNING: /workspace/TensorRT/t/oss-cicd/oss/parsers/onnx/onnx2trt_utils.cpp:227: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2021-06-18 01:43:24,479 [INFO] Extract_binaries for vad -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-voice-activity-detector-ctc-streaming/1
2021-06-18 01:43:24,480 [INFO] Extract_binaries for lm_decoder -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1
2021-06-18 01:43:24,510 [INFO] {'vocab_file': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/vocab.txt', 'decoding_language_model_binary': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/jarvis_asr_train_datasets_noSpgi_noLS_gt_3gram.binary', 'decoding_vocab': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/dict_vocab.txt', 'tokenizer_model': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/tokenizer.model'}
2021-06-18 01:43:24,510 [INFO] Model config has vocab file and tokenizer specified. Will create lexicon file from vocab_file /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/dict_vocab.txt and tokenizer model /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-ctc-decoder-cpu-streaming/1/tokenizer.model
2021-06-18 01:43:24,611 [INFO] processed 10000 lines
2021-06-18 01:43:24,714 [INFO] processed 20000 lines
2021-06-18 01:43:24,818 [INFO] processed 30000 lines
2021-06-18 01:43:24,923 [INFO] processed 40000 lines
2021-06-18 01:43:25,028 [INFO] processed 50000 lines
2021-06-18 01:43:25,132 [INFO] processed 60000 lines
2021-06-18 01:43:25,238 [INFO] processed 70000 lines
2021-06-18 01:43:25,343 [INFO] processed 80000 lines
2021-06-18 01:43:25,449 [INFO] processed 90000 lines
2021-06-18 01:43:25,554 [INFO] processed 100000 lines
2021-06-18 01:43:25,660 [INFO] processed 110000 lines
2021-06-18 01:43:25,765 [INFO] processed 120000 lines
2021-06-18 01:43:25,873 [INFO] processed 130000 lines
2021-06-18 01:43:25,980 [INFO] processed 140000 lines
2021-06-18 01:43:26,088 [INFO] processed 150000 lines
2021-06-18 01:43:26,195 [INFO] processed 160000 lines
2021-06-18 01:43:26,302 [INFO] processed 170000 lines
2021-06-18 01:43:26,407 [INFO] processed 180000 lines
2021-06-18 01:43:26,514 [INFO] processed 190000 lines
2021-06-18 01:43:26,621 [INFO] processed 200000 lines
2021-06-18 01:43:26,729 [INFO] processed 210000 lines
2021-06-18 01:43:26,836 [INFO] processed 220000 lines
2021-06-18 01:43:26,943 [INFO] processed 230000 lines
2021-06-18 01:43:26,971 [INFO] skipped 0 empty lines
2021-06-18 01:43:26,971 [INFO] filtered 0 lines
2021-06-18 01:43:26,973 [INFO] Extract_binaries for self -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming/1
2021-06-18 01:43:28,617 [INFO] Extract_binaries for featurizer -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-feature-extractor-streaming-offline/1
2021-06-18 01:43:28,618 [WARNING] /data/models/jarvis-trt-citrinet-1024 already exists, skipping deployment. To force deployment rerun with -f or remove the /data/models/jarvis-trt-citrinet-1024
2021-06-18 01:43:28,619 [INFO] Extract_binaries for vad -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-voice-activity-detector-ctc-streaming-offline/1
2021-06-18 01:43:28,620 [INFO] Extract_binaries for lm_decoder -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1
2021-06-18 01:43:28,649 [INFO] {'vocab_file': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/vocab.txt', 'decoding_language_model_binary': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/jarvis_asr_train_datasets_noSpgi_noLS_gt_3gram.binary', 'decoding_vocab': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/dict_vocab.txt', 'tokenizer_model': '/data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/tokenizer.model'}
2021-06-18 01:43:28,649 [INFO] Model config has vocab file and tokenizer specified. Will create lexicon file from vocab_file /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/dict_vocab.txt and tokenizer model /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline-ctc-decoder-cpu-streaming-offline/1/tokenizer.model
2021-06-18 01:43:28,750 [INFO] processed 10000 lines
2021-06-18 01:43:28,852 [INFO] processed 20000 lines
2021-06-18 01:43:28,956 [INFO] processed 30000 lines
2021-06-18 01:43:29,061 [INFO] processed 40000 lines
2021-06-18 01:43:29,166 [INFO] processed 50000 lines
2021-06-18 01:43:29,271 [INFO] processed 60000 lines
2021-06-18 01:43:29,376 [INFO] processed 70000 lines
2021-06-18 01:43:29,488 [INFO] processed 80000 lines
2021-06-18 01:43:29,597 [INFO] processed 90000 lines
2021-06-18 01:43:29,705 [INFO] processed 100000 lines
2021-06-18 01:43:29,808 [INFO] processed 110000 lines
2021-06-18 01:43:29,911 [INFO] processed 120000 lines
2021-06-18 01:43:30,021 [INFO] processed 130000 lines
2021-06-18 01:43:30,131 [INFO] processed 140000 lines
2021-06-18 01:43:30,237 [INFO] processed 150000 lines
2021-06-18 01:43:30,342 [INFO] processed 160000 lines
2021-06-18 01:43:30,447 [INFO] processed 170000 lines
2021-06-18 01:43:30,553 [INFO] processed 180000 lines
2021-06-18 01:43:30,667 [INFO] processed 190000 lines
2021-06-18 01:43:30,773 [INFO] processed 200000 lines
2021-06-18 01:43:30,878 [INFO] processed 210000 lines
2021-06-18 01:43:30,983 [INFO] processed 220000 lines
2021-06-18 01:43:31,088 [INFO] processed 230000 lines
2021-06-18 01:43:31,116 [INFO] skipped 0 empty lines
2021-06-18 01:43:31,116 [INFO] filtered 0 lines
2021-06-18 01:43:31,118 [INFO] Extract_binaries for self -> /data/models/citrinet-1024-asr-trt-ensemble-vad-streaming-offline/1
2021-06-18 01:43:32,013 [INFO] Extract_binaries for tokenizer -> /data/models/jarvis_tokenizer/1
2021-06-18 01:43:32,015 [INFO] Extract_binaries for language_model -> /data/models/jarvis-trt-jarvis_intent_weather-nn-bert-base-uncased/1
2021-06-18 01:43:35,045 [INFO] Building TRT engine from PyTorch Checkpoint
[TensorRT] ERROR: ../rtSafe/safeRuntime.cpp (25) - Cuda Error in allocate: 2 (out of memory)
[TensorRT] ERROR: ../rtSafe/safeRuntime.cpp (25) - Cuda Error in allocate: 2 (out of memory)
Traceback (most recent call last):
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/export_bert_pytorch_to_trt.py", line 976, in <module>
pytorch_to_trt()
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/export_bert_pytorch_to_trt.py", line 935, in pytorch_to_trt
return convert_pytorch_bert_to_trt(
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/export_bert_pytorch_to_trt.py", line 788, in convert_pytorch_bert_to_trt
with build_engine(
AttributeError: __enter__
2021-06-18 01:43:48,460 [ERROR] Traceback (most recent call last):
File "/opt/conda/lib/python3.8/site-packages/servicemaker/cli/deploy.py", line 87, in deploy_from_jmir
generator.serialize_to_disk(
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py", line 340, in serialize_to_disk
module.serialize_to_disk(repo_dir, jmir, config_only, verbose, overwrite)
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py", line 231, in serialize_to_disk
self.update_binary(version_dir, jmir, verbose)
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py", line 569, in update_binary
bindings = self.build_trt_engine_from_pytorch_bert(
File "/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py", line 532, in build_trt_engine_from_pytorch_bert
raise Exception("convert_pytorch_to_trt failed.")
Exception: convert_pytorch_to_trt failed.
+ echo
+ echo 'Jarvis initialization complete. Run ./jarvis_start.sh to launch services.'
Jarvis initialization complete. Run ./jarvis_start.sh to launch services.
I’m using a 3070 gpu which has 8GB I believe. The computer has 32GB of main memory. What should I do in order to run Jarvis?