404 Function not found for account when calling aisingapore/sea-lion-7b-instruct via integrate.api.nvidia.com

Hello NVIDIA Support,
I’m using the OpenAI-compatible endpoint:

  • Base URL: https://integrate.api.nvidia.com/v1

  • SDK: openai Python package

  • OS: Windows (Anaconda env)

Issue:
GET /v1/models lists aisingapore/sea-lion-7b-instruct, but POST /v1/chat/completions fails with 404 “Function not found”.

Repro code:

from openai import OpenAI
import os

client = OpenAI(
  base_url="https://integrate.api.nvidia.com/v1",
  api_key=os.environ["NVIDIA_API_KEY"],
)

# model appears in models.list()
models = client.models.list().data
print(any(m.id == "aisingapore/sea-lion-7b-instruct" for m in models))

# call fails
resp = client.chat.completions.create(
  model="aisingapore/sea-lion-7b-instruct",
  messages=[{"role":"user","content":"Ping"}],
  max_tokens=16,
  temperature=0,
)
print(resp.choices[0].message.content)

Error:

404 - "Function id '02f84bf4-c1a1-489b-a9de-ac3e8dcdec14' version 'null':
Specified function in account 'xBG7QnWfO9Ys1wHrinEcXIw9QhQMNJECT4TBke8wSBI' is not found"

Notes:

  • Other models from /v1/models work fine using the same key and endpoint.

  • Only aisingapore/sea-lion-7b-instruct fails.

Could you please check the entitlement/deployment mapping for this model on my account, or confirm if the model is currently unavailable for hosted inference?

Thanks!