If you are having memory issues, it may be an issue with your cache. Please reference our FAQ on how to clear the cache.
Related topics
| Topic | Replies | Views | Activity | |
|---|---|---|---|---|
| Qwen3-235B-A22B-NVFP4 Playbook Example Hangs | 5 | 502 | April 2, 2026 | |
| Qwen3.5-397B-A17B + DGX Spark (duo) | 56 | 5119 | April 13, 2026 | |
| DGX Spark crashes when running tensorrt-llm | 3 | 217 | March 7, 2026 | |
| Question on Inference Performance Results of Qwen3 235B A22B on 2× DGX Spark | 5 | 737 | December 19, 2025 | |
| [Bug] TensorRT-LLM 1.2.0rc8: "TRTLLMGenFusedMoE does not support SM120" error on DGX Spark with gpt-oss-120b + Eagle3 | 9 | 529 | February 17, 2026 | |
| [Issue] Qwen3-Next-80B NVFP4 and FP8 Cannot Be Served via trtllm-serve on DGX Spark GB10 (TRT-LLM 1.3.0rc7) | 2 | 212 | May 1, 2026 | |
| Bf16 LoRA Fine-Tuning of Qwen3.5-35B-A3B on DGX Spark — No Quantization Required | 5 | 768 | April 6, 2026 | |
| DGX Spark Multi-Node LLM Inference Report for Qwen3-235B model | 35 | 2211 | May 1, 2026 | |
| DGX Spark performance | 50 | 4821 | February 27, 2026 | |
| Qwen3.5-397B-A17B-int4-AutoRound - 4 x db10 node - updated results 37 - 94 tok/s | 26 | 1762 | April 28, 2026 |