I’m trying to run Llama 3.3 70B, and it cannot find a suitable profile using 2 or 3 H100 (80BG). I thought it would run using 2 H100:
container nvcr.io/nim/meta/llama-3.3-70b-instruct:1.5.2
I’ve been running other NIM llama 3.1 and 3.2 successfully.
Can I pass some options to make it run? fp8 or decreasing the context window?
Thanks,
Luc