Hi,
We are attempting to automate the lip sync pipeline using Nvidia ACE Audio2Face microservices.
We have implemented the animation pipeline workflow as suggested in NVIDIA Animation Pipeline with Omniverse Renderer ( Docker & Omniverse Animation Pipeline Workflow — ACE documentation) . The setup works with default-avatar-scene:1.0.0, but the lip sync quality is not great - Default avatar
However, when using our custom avatar, we followed all the required steps detailed in Custom Avatar Creation**. Despite this, the resulting custom avatar does not perform any lip sync when run through the same pipeline as above - Edited - custom avatar rendered with ACE pipeline
We are using audio generated using Azure TTS, 16 kHz sampling rate - [Audio_azure_tts](https ://drive.google.com/file/d/1GORk8gg9cwabCxBNJ9niu1nXKhuPO5E-/view?usp=sharing). We also tried with Nvidia’s TTS, the results were the same.
Would you be able to guide us on what could possibly be wrong?