niklasfrick

Real-time hardware and LLM inference monitoring — GPU, CPU, memory, and vLLM metrics streamed to a dashboard.

14
1
100% credibility
Found Apr 24, 2026 at 14 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
TypeScript
AI Summary

Spark Dashboard is a real-time web-based monitor for hardware resources and vLLM AI inference performance on Linux systems with NVIDIA GPUs.

How It Works

1
🔍 Discover Spark Dashboard

You hear about a handy tool to watch your computer's GPU and AI performance live on powerful Linux machines.

2
💻 Set it up on your GPU computer

Download and prepare the monitor on your Linux setup with NVIDIA graphics card—it takes just moments.

3
🚀 Start the live viewer

Turn on the service with one easy step, and it runs quietly in the background keeping an eye on everything.

4
🌐 Open in your web browser

Go to the web address shown, and instantly see colorful gauges and charts updating every second.

5
Check hardware health

Spot GPU temperature, power use, CPU load, and memory at a glance with heatmaps and trends.

6
🤖 Monitor your AI engines

Your running AI models appear automatically in tabs—watch speed, wait times, and batch sizes across many at once.

🎉 Perfect AI oversight

Enjoy smooth real-time insights, catch slowdowns early, and keep your AI projects humming happily.

Sign up to see the full architecture

5 more

Sign Up Free

Star Growth

See how this repo grew from 14 to 14 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is spark-dashboard?

Spark-dashboard is a real-time dashboard for monitoring hardware and LLM inference on Linux systems with NVIDIA GPUs. It polls GPU utilization, temperature, power, CPU cores, memory (including unified pools), disk, and network stats every second, while scraping vLLM metrics like tokens per second, TTFT, queue times, and KV cache usage via Prometheus. Built in Rust with a React frontend, it streams everything over WebSocket to a responsive UI with gauges, charts, and auto-rotating multi-engine tabs—install via `cargo install` and run as a systemd service on port 3000.

Why is it gaining traction?

In a sea of generic real time hardware monitors, this stands out with vLLM-specific smarts: auto-detects engines via process scan and Docker, aggregates across multiple instances in a "global" view, and flags GPU events like thermal throttling. The dev setup proxies to remote hosts seamlessly, and charts overlay inference requests on hardware traces for spotting bottlenecks. Early adopters on DGX Spark praise the glanceable metrics and 15-minute history without setup hassle.

Who should use this?

AI engineers tuning vLLM inference on NVIDIA workstations, DGX boxes, or cloud VMs—especially those juggling multi-engine setups for A/B testing models. It's ideal for real-time hardware in the loop debugging, like correlating batch sizes with power brakes during long runs. Skip if you're not on Linux/NVIDIA or need enterprise alerting.

Verdict

Grab it if vLLM monitoring is your pain point—solid for its niche despite 14 stars and 1.0% credibility score signaling early days. Docs and systemd integration are polished, but expect tweaks for production scale. Worth a spin over Grafana for quick local insights.

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.