Hello NVIDIA Support,
I’m using the OpenAI-compatible endpoint:
-
Base URL:
https://integrate.api.nvidia.com/v1 -
SDK:
openaiPython package -
OS: Windows (Anaconda env)
Issue:
GET /v1/models lists aisingapore/sea-lion-7b-instruct, but POST /v1/chat/completions fails with 404 “Function not found”.
Repro code:
from openai import OpenAI
import os
client = OpenAI(
base_url="https://integrate.api.nvidia.com/v1",
api_key=os.environ["NVIDIA_API_KEY"],
)
# model appears in models.list()
models = client.models.list().data
print(any(m.id == "aisingapore/sea-lion-7b-instruct" for m in models))
# call fails
resp = client.chat.completions.create(
model="aisingapore/sea-lion-7b-instruct",
messages=[{"role":"user","content":"Ping"}],
max_tokens=16,
temperature=0,
)
print(resp.choices[0].message.content)
Error:
404 - "Function id '02f84bf4-c1a1-489b-a9de-ac3e8dcdec14' version 'null':
Specified function in account 'xBG7QnWfO9Ys1wHrinEcXIw9QhQMNJECT4TBke8wSBI' is not found"
Notes:
-
Other models from
/v1/modelswork fine using the same key and endpoint. -
Only
aisingapore/sea-lion-7b-instructfails.
Could you please check the entitlement/deployment mapping for this model on my account, or confirm if the model is currently unavailable for hosted inference?
Thanks!