I have been trying to reconstruct the 3-D scenes using multiple monocular images of scene captured from different positions/angles (tried with video frames also). I am aiming to generate the usdz file which I can use with the Nvidia NuRec (Carla + Nurec).
Here are the articles I have followed:
Sharing some of my outputs.
Please suggest how can I improve my reconstruction quality to achieve same quality output as that of examples present in Nvidia Physical AI repository. Also if there are other ways to generate the same (For example Depth Anything 3 also offers 3-d reconstruction and Gaussian Splatting based novel view which seems to provide better reconstruction but no integration with NuRec).
Hi and thanks for posting the question. You are certainly getting the hang of it. The second one, with the office chair looks great. For tips you just need to stick to the recommendations given right there in the “Workflow” section, like:
Ensure steady lighting and proper focus
Avoid fast motion and blur
Use faster shutter speed (1/100 second or faster recommended)
Lock focus and exposure to maintain consistency across shots
Thanks for the advice. Issue seems to be from colmap side tho I am not sure why. I was using colmap with GUI and results were bad but using the cli commands with same input images I got much much better results. Thanks for the support.
Hi @Richard3D I was able to generate the USDZ file. I want to run my custom USDZ file for CARLA + NuRec rendering as of now I have tried using the USDZ output of 3dgrut but it is not compatible as the files inside the USDZ are different (3dgrut USDZ vs Nvidia Physical AI repository). Can you help me guiding how can I generate USDZ file compatible for Carla + NuRec rendering.
I have run reconstruction the reconstruction on KITTI dataset video and the construction but when trying to get a novel view I am seeing noise/artifacts. Please suggest how can I improve Novel view synthesis without artifacts.
Looks really good. I don’t think you can get much better than that at the current level of technology. It is just a question of working with the limits of the NuRec and GS. Again, the only thing you can do to improve is more and higher quality input imagery. This is not really an Omniverse issue. This is just the current processing limit of the workflow outside of Nvidia.