⚡ Native MLX Swift LLM inference server for Apple Silicon. OpenAI-compatible API, SSD streaming for 100B+ MoE models, TurboQuant KV cache compression, + iOS iPhone app.
SwiftLM is a native Swift server for running large language models on Apple Silicon Macs with an OpenAI-compatible API and an optional iOS chat app.
How It Works
You hear about a super-fast way to run powerful AI chatbots right on your Mac without needing the internet.
Download the simple app file from the releases page and unzip it on your Mac.
Open your terminal, type the app name with a model like a smart assistant, and it grabs everything needed.
Ask questions using a web tool or app, and get instant smart replies powered by your Mac's chip.
Keep enjoying fast chats on your computer.
Run the same AI on your iPhone or iPad.
You now have a speedy, private thinking assistant that works offline on your Apple devices.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.