noonghunna / club-3090
PublicCommunity recipes for serving LLMs on RTX 3090. Multi-engine (vLLM, llama.cpp, SGLang) and model-agnostic. Currently shipping Qwen3.6-27B configs for 1× and 2× cards.
A collection of validated configurations and recipes for running large vision-and-tools-enabled language models locally on one or two RTX 3090 GPUs via OpenAI-compatible servers.
How It Works
You hear about a way to run powerful AI chatbots right on your gaming computer with RTX 3090 graphics cards, without needing the internet.
Download the simple recipe folder that has everything you need to get started at home.
Fetch the smart AI model files (about 20GB) that your computer will use to think and chat.
Perfect for one RTX 3090 – quick setup for personal chatting and tools.
Use two RTX 3090s for faster responses and handling bigger conversations.
Start everything with one easy command – watch it boot up in a couple minutes and get ready to chat.
Send a message like 'What's the capital of France?' and see it reply instantly, just like online services.
Now you have a full-featured AI assistant at home – chat, analyze images, use tools, all fast and private on your own hardware.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.