jdaln / dgx-spark-inference-stack
PublicServe the home! Inference stack for your Nvidia DGX Spark aka the Grace Blackwell AI supercomputer on your desk. Mostly vLLM based for now
A ready-to-use home server kit for Nvidia DGX Spark that runs many large AI language models on demand with smart power saving and easy app connections.
How It Works
You find a friendly guide to turn your powerful Nvidia DGX Spark computer into a smart home helper for chatting with AI anytime.
Create a few folders on your computer and download special word lists needed for the AI to understand language.
Make a free account at a helpful service and prepare your computer tools so everything connects smoothly.
Prepare custom smart engines tailored for your computer, taking about 20 minutes each to ensure top speed and smarts.
Start your home AI server – it loads brains only when needed and saves power by resting when idle.
Send a simple hello message and get a friendly reply, proving your setup works perfectly.
Now enjoy powerful conversations, coding help, or image analysis anytime from apps like your code editor or terminal, all from your own machine.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.