Hi NV member:
I would like to use the script from "GitHub - eugr/spark-vllm-docker: Docker configuration for running VLLM on dual DGX Sparks
" to quickly set up a GPT OSS 120B server so that OpenClaw can access this server to run related services. However, after following the steps below, I encounter a “model not found” issue. Could this be caused by some incompatibility in vLLM?
Have the following error happen when chat in openclaw ui
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-11 08:06:30 [protocol.py:117] The following fields were present in the request but ignored: {‘store’}
(APIServer pid=176) ERROR 02-11 08:06:30 [serving_chat.py:236] Error with model error=ErrorInfo(message=‘The model gpt-oss-120b does not exist.’, type=‘NotFoundError’, param=‘model’, code=404)
(APIServer pid=176) INFO: 127.0.0.1:47302 - “POST /v1/chat/completions HTTP/1.1” 404 Not Found
I have no experience with OpenClaw, and I cannot be more precise, unfortunately.
That being said, the APIServer logs suggest a model Id mismatch – I may be mistaken.
Now, from your OpenClaw configuration it is not apparent the presence of wrong model identifiers. Hence I wonder whether further identifiers are present elsewhere in the system configuration or code.
Hi @Keyper-AI and @eugr :
After adding the --served-model-name gpt-oss-120b argument and updating the OpenClaw ID, I no longer see the “gpt-oss-120b” failure. However, the vLLM server still shows a 400 Bad Request error.
I tested using Ollama with gpt-oss-120b, and it works normally. Is there any specific configuration I should double-check or pay attention to?
Thank you for your help.
Add the --served-model-name gpt-oss-120b to spark-vllm-docker
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=176) WARNING 02-12 02:05:35 [protocol.py:117] The following fields were present in the request but ignored: {‘store’}
(APIServer pid=176) INFO: 127.0.0.1:60938 - “POST /v1/chat/completions HTTP/1.1” 400 Bad Request
Hi @Keyper-AI, @eugr and @adg1
It is now working normally with OpenClaw in the Spark-VLLM-Docker + GPT-OSS-120B environment after applying the following steps and modifications. Thank you for your helpful suggestions and methods.
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘strict’}
(APIServer pid=175) WARNING 02-12 06:09:14 [protocol.py:117] The following fields were present in the request but ignored: {‘store’}
(APIServer pid=175) INFO: 127.0.0.1:47234 - “POST /v1/chat/completions HTTP/1.1” 200 OK
(APIServer pid=175) INFO 02-12 06:09:16 [loggers.py:257] Engine 000: Avg prompt throughput: 2138.2 tokens/s, Avg generation throughput: 51.4 tokens/s, Running: 1 reqs, Waiting: 0 reqs, GPU KV cache usage: 2.2%, Prefix cache hit rate: 80.3%
(APIServer pid=175) INFO 02-12 06:09:26 [loggers.py:257] Engine 000: Avg prompt throughput: 0.0 tokens/s, Avg generation throughput: 11.0 tokens/s, Running: 0 reqs, Waiting: 0 reqs, GPU KV cache usage: 0.0%, Prefix cache hit rate: 80.3%
(APIServer pid=175) INFO 02-12 06:09:36 [loggers.py:257] Engine 000: Avg prompt throughput: 0.0 tokens/s, Avg generation throughput: 0.0 tokens/s, Running: 0 reqs, Waiting: 0 reqs, GPU KV cache usage: 0.0%, Prefix cache hit rate: 80.3%
Hi @eugr :
As you mentioned, when my chat gets longer and longer, perhaps because the context becomes larger and larger, after it is sent to the gpt-oss-120b server, the following error keeps occurring. I assume every model has this issue, right? Or based on your understanding, are there better models or approaches to handle this?
Thank you.
Error message:
400 max_tokens must be at least 1, got -17474. (parameter=max_tokens, value=-17474)
Don’t set “maxTokens”: 8192, it’s too low and restricts model to output anything longer than that. Just don’t set it at all. Also reduce your context window parameter in openclaw settings to 131072 - maximum that gpt-oss-120b supports.
Hi @eugr :
Thank you for your suggestion. You’re absolutely right. As you mentioned, the context window should better align with GPT-OSS-120B.
We can experiment later with OpenClaw plus multiple agents (Claude Code), combined with either a single or multiple DGX Spark systems. That would also be an interesting application scenario for DGX Spark edge devices.