ellipticmarketing

Local router that benchmarks free coding models across providers and forwards requests to the best available model. Compatible with Opencode and Openclaw

79
7
100% credibility
Found Feb 26, 2026 at 32 stars 2x -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
JavaScript
AI Summary

Local router that benchmarks and routes requests to the fastest available free coding AI models from multiple providers via an OpenAI-compatible endpoint.

How It Works

1
🔍 Discover ModelRelay

You hear about this smart tool that automatically picks the quickest free AI brain for helping with coding tasks.

2
💻 Install easily

With one simple command, you add it to your computer and it's ready to go.

3
🔗 Connect AI services

You link your free accounts from helpful places like Groq or NVIDIA, so the tool knows where to send questions.

4
🚀 Start the magic

You launch it once, and it hums quietly in the background, always watching for the fastest responder.

5
📊 Check the dashboard

You open a friendly screen showing all the AIs, their speeds, who's winning, and everything at a glance.

6
⚙️ Hook up your coding app

You point your favorite coding companion to this tool, and it starts using the best AI automatically.

🎉 Code with lightning speed

Now your coding gets supercharged help from the fastest free AI every time, feeling effortless and fun.

Sign up to see the full architecture

5 more

Sign Up Free

Star Growth

See how this repo grew from 32 to 79 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is modelrelay?

modelrelay is a JavaScript CLI tool that spins up a local OpenAI-compatible router, pinging and benchmarking free coding models across providers like NVIDIA NIM, Groq, Cerebras, OpenRouter, and more. Point your requests at http://127.0.0.1:7352/v1 with model "auto-fastest," and it routes to the lowest-latency option available. Install via npm, onboard API keys with one command, and get a web dashboard for real-time stats—perfect for dodging provider outages without changing your code.

Why is it gaining traction?

It auto-selects the fastest model via continuous latency checks, with proactive retries across providers and a ban list for flaky ones. Drop-in compatibility with OpenCode and OpenClaw (auto-configures their JSON), plus autostart on login and request logging, makes setup effortless. As a local GitHub Copilot alternative, it proxies free inference without cloud dependencies, hooking devs chasing peak coding AI speed.

Who should use this?

Coders in OpenCode or OpenClaw needing a local router URL to aggregate free models for reliable completions. Full-stack teams building local GitHub actions runners or local GitHub Copilot alternatives on self-hosted setups. Devs tweaking local router config for shared machines, dodging slow APIs across providers.

Verdict

Worth npm installing for free-model routing if you hit latency walls—CLI shines, docs cover OpenCode integration cleanly (28 stars). 1.0% credibility score and low stars mean it's early; prod users should monitor stability, but the auto-benchmarking dashboard delivers immediate value.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.