|
vLLM container out of date for new models
|
10
|
1568
|
November 14, 2025
|
|
Install and Use vLLM for Inference on two Sparks does not work
|
159
|
3765
|
December 9, 2025
|
|
vLLM on GB10: gpt-oss-120b MXFP4 slower than SGLang/llama.cpp... what’s missing?
|
135
|
3802
|
February 14, 2026
|
|
New pre-built vLLM Docker Images for NVIDIA DGX Spark
|
48
|
2697
|
February 13, 2026
|
|
GLM-4.7-Flash-NVFP4 was just released, but for Transformers 5.0 + vLLM 0.14...?
|
89
|
3175
|
February 13, 2026
|
|
I'd like to learn how to use the latest vLLM on DGX Spark
|
9
|
1651
|
November 29, 2025
|
|
Run VLLM in Thor from VLLM Repository
|
15
|
1502
|
November 29, 2025
|
|
New bleeding-edge vLLM Docker Image: avarok/vllm-nvfp4-gb10-sm120
|
35
|
1592
|
December 31, 2025
|
|
Help: Running NVFP4 model on 2x DGX Spark with vLLM + Ray (multi-node)
|
18
|
1559
|
December 25, 2025
|
|
Setting up vLLM, SGLang or TensorRT on two DGX Sparks
|
16
|
1016
|
December 7, 2025
|