LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the macOS menu bar
oMLX is a menu bar application for running large language models locally on Apple Silicon Macs, featuring an admin dashboard for model downloads, multi-model serving, and OpenAI/Anthropic API compatibility.
How It Works
You hear about a simple app that lets your Mac run smart AI helpers right at home, without needing the internet.
Grab the ready-to-use app from the official releases page and drag it to your Applications folder.
Open the app from your menu bar, pick a folder for your AI helpers, and download your first one with a few clicks.
Jump into the built-in chat window to ask questions and get instant smart replies from your local AI.
Link it to your favorite writing apps or coding helpers so they can use your speedy local brain.
Watch your Mac handle big conversations smoothly from the menu bar, with everything private and blazing fast.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.