Any improvements in performance for gpt-oss-120b?
Related topics
| Topic | Replies | Views | Activity | |
|---|---|---|---|---|
| vLLM container out of date for new models | 10 | 1407 | November 14, 2025 | |
| Install and Use vLLM for Inference on two Sparks does not work | 159 | 3216 | December 9, 2025 | |
| I'd like to learn how to use the latest vLLM on DGX Spark | 9 | 1401 | November 29, 2025 | |
| Run VLLM in Thor from VLLM Repository | 15 | 1286 | November 29, 2025 | |
| vLLM on GB10: gpt-oss-120b MXFP4 slower than SGLang/llama.cpp... what’s missing? | 38 | 1516 | January 23, 2026 | |
| New bleeding-edge vLLM Docker Image: avarok/vllm-nvfp4-gb10-sm120 | 35 | 1140 | December 31, 2025 | |
| Help: Running NVFP4 model on 2x DGX Spark with vLLM + Ray (multi-node) | 18 | 1214 | December 25, 2025 | |
| Setting up vLLM, SGLang or TensorRT on two DGX Sparks | 16 | 782 | December 7, 2025 | |
| vLLM container 25.10-py3 fails to start | 13 | 458 | December 8, 2025 | |
| Announcing new VLLM container & 3.5X increase in Gen AI Performance in just 5 weeks of Jetson AGX Thor Launch | 46 | 2877 | December 14, 2025 |