6/4/2025
Updated with local NIM support via compose feature in AI Workbench
6/4/2025
Updated with local NIM support via compose feature in AI Workbench
08/04 - Update default NV-Embed-QA embedding model endpoint to nvidia/nv-embedqa-e5-v5
When try to upload files or urls for RAG receive this error:
âDECISION: GENERATION ADDRESSES QUESTIONâ
Failed to send telemetry event ClientStartEvent: capture() takes 1 positional argument but 3 were given
Failed to send telemetry event ClientCreateCollectionEvent: capture() takes 1 positional argument but 3 were given
[upload] Vectorstore creation failed: [504] Gateway Timeout
{â_contentâ: bâ', â_content_consumedâ: True, â_nextâ: None, âstatus_codeâ: 504, âheadersâ: {âDateâ: âTue, 05 Aug 2025 16:41:40 GMTâ, âTransfer-Encodingâ: âchunkedâ, âConnectionâ: âkeep-aliveâ, âAccess-Control-Allow-Credentialsâ: âtrueâ, âAccess-Control-Expose-Headersâ: ânvcf-reqidâ, âNvcf-Reqidâ: â2aa7a6bf-8182-4375-bb83-49f50be9a389â, âNvcf-Statusâ: âerroredâ, âVaryâ: âOrigin, origin, access-control-request-method, access-control-request-headersâ}, ârawâ: <urllib3.response.HTTPResponse object at 0x74110702bd30>, âurlâ: âhttps://ai.api.nvidia.com/v1/retrieval/nvidia/embeddingsâ, âencodingâ: None, âhistoryâ: , âreasonâ: âGateway Timeoutâ, âcookiesâ: <RequestsCookieJar>, âelapsedâ: datetime.timedelta(seconds=302, microseconds=543075), ârequestâ: <PreparedRequest [POST]>, âconnectionâ: <requests.adapters.HTTPAdapter object at 0x7410fcb1f6a0>}
The https://ai.api.nvidia.com appears dont exist anymore
Any suggestion?
Thanks1
08/04 - Update default
NV-Embed-QAembedding model endpoint tonvidia/nv-embedqa-e5-v5
Looks like the default embedding model endpoint has been removed. The project has since been updated to a new embedding model endpoint. Can you pull down the most recent changes and try again? Thanks!
Hi,
Iâm having an error in the initial build process
[16/19] RUN pip install --user -r /opt/project/build/requirements.txt:
836.1 status = run_func(*args)
836.1 File â/usr/lib/python3/dist-packages/pip/_internal/cli/req_command.pyâ, line 205, in wrapper
836.1 return func(self, options, args)
836.1 File â/usr/lib/python3/dist-packages/pip/_internal/commands/install.pyâ, line 389, in run
836.1 to_install = resolver.get_installation_order(requirement_set)
836.1 File â/usr/lib/python3/dist-packages/pip/_internal/resolution/resolvelib/resolver.pyâ, line 188, in get_installation_order
836.1 weights = get_topological_weights(
836.1 File â/usr/lib/python3/dist-packages/pip/_internal/resolution/resolvelib/resolver.pyâ, line 276, in get_topological_weights
836.1 assert len(weights) == expected_node_count
836.1 AssertionError
------
ERROR: failed to build: failed to solve: process â/bin/bash -c pip install --user -r /opt/project/build/requirements.txtâ did not complete successfully: exit code: 2
Iâm have the same error.
[16/19] RUN pip install --user -r /opt/project/build/requirements.txt:
44.50 weights = get_topological_weights(
44.50 File â/usr/lib/python3/dist-packages/pip/_internal/resolution/resolvelib/resolver.pyâ, line 276, in get_topological_weights
44.50 assert len(weights) == expected_node_count
44.50 AssertionError
------
Containerfile:61
--------------------
60 |
61 | >>> RUN pip install --user \
62 | >>> -r /opt/project/build/requirements.txt
63 |
--------------------
ERROR: failed to build: failed to solve: process â/bin/bash -c pip install --user -r /opt/project/build/requirements.txtâ did not complete successfully: exit code: 2
The README doesnât describe how the two different endpoints are used. I looked through the agentic_rag docs, but donât see it. There is a diagram for the intermediate mode, but it doesnât tell me which conceptual model is used at any of the stages.
NVIDIA_API_KEYTAVILY_API_KEYThis comment
Same error as shown by other above.
Host system: Mac
Hi
Same issue on DGX Spark :
[16/19] RUN pip install --user -r /opt/project/build/requirements.txt:
44.50 weights = get_topological_weights(
44.50 File â/usr/lib/python3/dist-packages/pip/_internal/resolution/resolvelib/resolver.pyâ, line 276, in get_topological_weights
44.50 assert len(weights) == expected_node_count
44.50 AssertionError
------
Containerfile:61
--------------------
60 |
61 | >>> RUN pip install --user \
62 | >>> -r /opt/project/build/requirements.txt
63 |
--------------------
ERROR: failed to build: failed to solve: process â/bin/bash -c pip install --user -r /opt/project/build/requirements.txtâ did not complete successfully: exit code: 2
==> SOLVED BY :
Adding those lines into preBuild.bash file : edit from AI Workbench console
pip install --upgrade pip
If not enough :
add in first line into requirements.txt
pip>=23.3
=> ISSUE SOLVED
I hope the same for you !
Inserting the line above as the first line in preBuild.bash fixed for me.
I updated github issue as well in case you are following the instructions on DGX Spark | Try NVIDIA NIM APIs
(11/04) - Bump pip version and base container version tag.
Thanks for the catch, all. Looks like an out of date pip manager version. Simple version bump has been pushed.
iâve tried everything and iâm still getting 403 error.
used echo $(EACH KEY) in cli to confirm keys are set, ânvapiâŠâ. double checked nvidia key is set for public endpoints. generated a second key. i tried generating the key for specific model. restarted container and rebuilt it. then recloned git and did everything again.
but still 403 for me. any ideas out there?
File â/home/workbench/.local/lib/python3.10/site-packages/langchain_nvidia_ai_endpoints/_common.pyâ, line 462, in _try_raise
raise Exception(f"{header}\\n{body}") from None
Exception: [403] Forbidden
Authorization failed
Hi, a 403 message typically means you have logged in properly via the API key (eg. it is a valid key), but that the permissions set for that key are improper.
Ensure you had included the âPublic API Endpointsâ permissions when you generated the key if done on NGC.
If generated from a model card on build.nvidia.com, ensure the selected Org visible under your profile dropdown menu has public API Key access (badge should show green with a checkmark instead of gray).
01/15/2026
Updated model selection to include Qwen 253B
Updated Llama 70B from 3.1 to 3.3
Updated documents web links list to reflect the current AI Workbench documentation sites