Please provide the following information when requesting support.
Hardware - GPU (T4)
Hardware - CPU
Operating System - ubuntu
Riva Version = 1.0 Beta
TLT Version (if relevant)
How to reproduce the issue ? (This is for errors. Please share the command and the detailed log here)
Fine tune quartznet model with custom language model. it succeeds
riva-build speech_recognition
/servicemaker-dev/riva-custom-speech.rmir:tlt_encode
/servicemaker-dev/quartznet_asr_set_1pt2.riva:tlt_encode
–name riva-custom-speech-quartznet
–decoder_type=os2s
–decoding_language_model_binary=/lm/text.binary -f
2021-08-21 18:01:26,090 [INFO] Packing binaries for nn
2021-08-21 18:01:26,663 [INFO] Trying to extract from model quartznet_asr_set_1pt2.riva
2021-08-21 18:01:27,207 [INFO] Packing binaries for lm_decoder
2021-08-21 18:01:27,207 [INFO] Trying to copy model binary from /tmp/tmpd4t7f589/vocab.txt into rmir at /servicemaker-dev/riva-custom-speech.rmir.
2021-08-21 18:01:27,208 [INFO] Trying to copy model binary from /lm/text.binary into rmir at /servicemaker-dev/riva-custom-speech.rmir.
2021-08-21 18:01:27,209 [INFO] Packing binaries for rescorer
2021-08-21 18:01:27,210 [INFO] Trying to copy model binary from /tmp/tmpd4t7f589/vocab.txt into rmir at /servicemaker-dev/riva-custom-speech.rmir.
2021-08-21 18:01:27,210 [INFO] Packing binaries for vad
2021-08-21 18:01:27,211 [INFO] Trying to copy model binary from /tmp/tmpd4t7f589/vocab.txt into rmir at /servicemaker-dev/riva-custom-speech.rmir.
---------------------- SUCESS SO FAR--------
but when trying to do riva_init.sh i get following error
2021-08-21 18:22:08,719 [INFO] Building TRT engine from ONNX file
[libprotobuf WARNING /workspace/TensorRT/t/oss-cicd/oss/build/third_party.protobuf/src/third_party.protobuf/src/google/protobuf/io/coded_stream.cc:604] Reading dangerously large protocol message. If the message turns out to be larger than 2147483647 bytes, parsing will be halted for security reasons. To increase the limit (or to disable these warnings), see CodedInputStream::SetTotalBytesLimit() in google/protobuf/io/coded_stream.h.
[libprotobuf WARNING /workspace/TensorRT/t/oss-cicd/oss/build/third_party.protobuf/src/third_party.protobuf/src/google/protobuf/io/coded_stream.cc:81] The total number of bytes read was 564429124
[TensorRT] WARNING: /workspace/TensorRT/t/oss-cicd/oss/parsers/onnx/onnx2trt_utils.cpp:227: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
2021-08-21 18:31:48,833 [INFO] Extract_binaries for vad → /data/models/riva-custom-speech-voice-activity-detector-ctc-streaming/1
2021-08-21 18:31:48,835 [INFO] Extract_binaries for lm_decoder → /data/models/riva-custom-speech-ctc-decoder-cpu-streaming/1
2021-08-21 18:31:48,837 [ERROR] Couldn’t extract /lm/vocab.txt to /data/models/riva-custom-speech-ctc-decoder-cpu-streaming/1
Traceback (most recent call last):
File “/opt/conda/lib/python3.8/site-packages/servicemaker/rmir/rmir.py”, line 314, in extract
raise ValueError(f"{target_name} not in artifacts")
ValueError: lm_decoder-vocab.txt not in artifacts
2021-08-21 18:31:48,838 [INFO] {‘vocab_file’: ‘/data/models/riva-custom-speech-ctc-decoder-cpu-streaming/1/vocab.txt’, ‘decoding_language_model_binary’: ‘/data/models/riva-custom-speech-ctc-decoder-cpu-streaming/1/text.binary’, ‘tokenizer_model’: ‘/data/models/riva-custom-speech-ctc-decoder-cpu-streaming/1/tokenizer.model’}
2021-08-21 18:31:48,838 [ERROR] Traceback (most recent call last):
File “/opt/conda/lib/python3.8/site-packages/servicemaker/cli/deploy.py”, line 85, in deploy_from_rmir
generator.serialize_to_disk(
File “/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py”, line 340, in serialize_to_disk
module.serialize_to_disk(repo_dir, rmir, config_only, verbose, overwrite)
File “/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py”, line 231, in serialize_to_disk
self.update_binary(version_dir, rmir, verbose)
File “/opt/conda/lib/python3.8/site-packages/servicemaker/triton/asr.py”, line 409, in update_binary
vocab_file = RivaConfigGenerator.get_binary_from_key(self, copied, ‘decoding_vocab’)
File “/opt/conda/lib/python3.8/site-packages/servicemaker/triton/triton.py”, line 294, in get_binary_from_key
raise Exception(binary_key + " not found in .rmir")
Exception: decoding_vocab not found in .rmir
-
echo
-
echo ‘Riva initialization complete. Run ./riva_start.sh to launch services.’
Riva initialization complete. Run ./riva_start.sh to launch services.