Llama-3.1-Nemotron-Ultra-253B-v1 API not working

It’s been saying ‘Internal server error’ since Saturday. Can we look into this? Thanks.

Thanks for bringing this to our attention - I’ve asked the team to redeploy the endpoint.

Sophie

1 Like

It’s now back up and running - thanks again for letting us know it wasn’t working @vutran_kpt

Sophie

1 Like

Thank you for fixing it. Can we also look into Llama-3.3-Nemotron-Super-49B-v1.5? If we disable reasoning, the model shows ‘ErrorEngineCore encountered an issue. See stack trace (above) for the root cause.’ Thanks.

Hi @vutran_kpt,

I’m struggling to recreate the error you’re seeing with the llama-3_3-nemotron-super-49b-v1_5 model - are you using it through the web UI or deploying elsewhere?

Thanks,

Sophie


The Web UI shows this error when using the model with reasoning off. Same with API.

Interesting! I was asking the model the suggested math question, rather than ‘counting the r’s in strawberry’ question - it works fine on the maths!

I’ll let the team know and get it sorted!

Sophie

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.