I did a hosted compatibility repro against https://integrate.api.nvidia.com/v1 and put the artifacts here:
The question is narrower than “Codex support” or “NIM support” generally. I’m trying to clarify the current hosted Integrate contract for /v1/responses.
Exact results:
- direct
GET /v1/modelsreturned200 - direct
POST /v1/chat/completionsreturned200 - direct
POST /v1/responsesreturned404 page not found - the same
/v1/responses404also appeared in a widened six-model matrix on the same hosted Integrate surface across NVIDIA, Meta, and Mistral instruct models:nvidia/nemotron-3-super-120b-a12bnvidia/llama-3.3-nemotron-super-49b-v1nvidia/llama-3.1-nemotron-70b-instructnvidia/nemotron-mini-4b-instructmeta/llama-3.3-70b-instructmistralai/mistral-large
- NVIDIA’s current LLM NIM release notes describe experimental
ResponsesAPI support, which is why I’m asking whether this hosted result is expected
The narrow question is:
- is
/v1/responsesexpected to work today on the hosted Integrate surface? - if yes, is it limited to a different subset of models, accounts, or endpoint variants than the ones tested here?
If it helps, I can also provide the exact per-model request/response captures from the repro repo artifacts.