antirez / llama.cpp-deepseek-v4-flash
PublicExperimental implementation of DeepSeek v4 flaash in llama.cpp
Fork of llama.cpp adding experimental support for quantized DeepSeek v4 Flash models optimized for Apple Silicon.
How It Works
You hear about a way to run cutting-edge AI conversations right on your MacBook without needing the internet.
Download the special AI file from a safe sharing site with one click.
Get the free runner tool and open it up.
Type your first question and watch the AI respond like a frontier expert, thinking deeply just for you.
Chat endlessly with top-tier intelligence on your own MacBook, fast and secure.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.