|
MSI EdgeXpert Suddenly Power-Off During llama-benchy – Possible PD Firmware Issue?
|
|
0
|
4
|
March 5, 2026
|
|
HOW-TO: Run Qwen3-Coder-Next on Spark
|
|
88
|
5784
|
March 5, 2026
|
|
Run Qwen3.5-27B with spark-vllm-docker
|
|
1
|
15
|
March 5, 2026
|
|
DGX Spark + Qwen3-Next-80B: Proven Performance, But Missing Clear Path to NIM, TensorRT-LLM & Web UIs
|
|
14
|
2787
|
March 5, 2026
|
|
DGX Spark Performance Degradation - GPU Power Draw Issue
|
|
8
|
432
|
March 4, 2026
|
|
Guide: llama.cpp + Qwen3.5-35B-A3B + openclaw on GB10
|
|
4
|
2214
|
March 4, 2026
|
|
QMD + node-llama-cpp on Jetson Orin AGX: GPU runtime OOM
|
|
3
|
49
|
March 4, 2026
|
|
"unable to allocate CUDA0 buffer" after Updating Ubuntu Packages
|
|
241
|
13451
|
March 4, 2026
|
|
Step-3.5-Flash on Single Spark with 256k context
|
|
2
|
196
|
March 3, 2026
|
|
LLM library recomendations for maximum token speeds
|
|
9
|
202
|
March 2, 2026
|
|
Bypassing Python: Piping local LLM inference directly into a deterministic C++ compiler pipeline
|
|
2
|
23
|
March 1, 2026
|
|
Llama.cpp rpc on dgx spark
|
|
4
|
202
|
March 1, 2026
|
|
Building Local + Hybrid LLMs on DGX Spark That Outperform Top Cloud Models
|
|
14
|
2067
|
March 1, 2026
|
|
Web GUI Manager Easy Inference for the Spark
|
|
1
|
22
|
February 28, 2026
|
|
VLLM -- the $150M train wreck?
|
|
24
|
743
|
February 27, 2026
|
|
Investigating 513MHz cap for GPU
|
|
5
|
278
|
February 26, 2026
|
|
Nvidia spark dgx GB10 fine-tune slow time problem - Urgent HELP
|
|
5
|
33
|
February 26, 2026
|
|
Moving from Mac to NVIDIA: bought powerful hardware, but drowning in configs
|
|
37
|
1723
|
February 25, 2026
|
|
Building llama.cpp container images for Spark/GB10
|
|
15
|
1620
|
February 24, 2026
|
|
Llama.cpp on Jetson Orin NX 16GB for API-Only Inference — Bare Metal or NVIDIA Docker?
|
|
1
|
70
|
February 24, 2026
|
|
VSS Jetson Thor: GPU memory increase during summarization causes OOM unless VSS is restarted
|
|
1
|
24
|
February 24, 2026
|
|
Why is nvidia-uvm required for cuda
|
|
3
|
114
|
February 23, 2026
|
|
Qwen3-Next AWQ 4bit vs FP8 vs NVFP4 on single spark
|
|
7
|
823
|
February 23, 2026
|
|
(sparkrun) Qwen3.5 GGUF Benchmarks over llama.cpp RPC
|
|
1
|
340
|
February 20, 2026
|
|
MiniMax-2.5 on DGX Spark (thanks to Unsloth https://unsloth.ai/docs/models/minimax-2.5)
|
|
12
|
2289
|
February 20, 2026
|
|
Did the latest update brick my Spark?
|
|
2
|
222
|
February 19, 2026
|
|
How can i integrate NVIDIA LLM's to my system?
|
|
0
|
39
|
February 19, 2026
|
|
Best Inference Framework & Open Models for Orchestrator-Workers Agentic Coding on GB10 + 5090 Hybrid?
|
|
1
|
349
|
February 19, 2026
|
|
Llama.cpp GLM 4.7 Flash Benchmark
|
|
1
|
309
|
February 19, 2026
|
|
DGX Spark 64k kernels
|
|
2
|
222
|
February 19, 2026
|