Ashfaqbs

a collection of tiny llms with usecases

15
1
100% credibility
Found Mar 14, 2026 at 15 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

This repository offers demo scripts and servers for three small open-source language models that handle tool calling for tasks like weather lookups, calculations, and contact searches using local inference.

How It Works

1
🔍 Discover Smart Helpers

You stumble upon this collection of tiny AI demos that let you run smart assistants right on your own computer for free.

2
📥 Get the Free Runner

Download and set up a simple free app that powers these small AIs on your machine without needing the internet.

3
🧠 Download Tiny AI Brains

Grab three lightweight AI models – a quick router, a nano thinker, and a full agent – each under a few gigabytes.

4
Pick a Demo Folder

Choose one ready-made example folder to start experimenting with tool-using AI.

5
Choose Your Play Style
▶️
Quick Test

Run simple questions directly and see instant results.

🌐
Web Service

Start a local web spot to send questions and get smart replies.

6
💬 Ask Everyday Questions

Type in queries like 'weather in Tokyo' or 'add 25 and 17', and watch the AI pick the right action and give answers.

🎉 AI Magic at Home

Celebrate having your own fast, private smart helper that saves money and works offline for routing tasks and simple smarts.

Sign up to see the full architecture

5 more

Sign Up Free

Star Growth

See how this repo grew from 15 to 15 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is TinyLLM-usecases?

This Python repo delivers ready-to-run demos for three tiny LLMs—FunctionGemma 270M, Qwen3 0.6B, and Qwen3 4B—using Ollama and LangChain to handle tool calling, like routing "weather in Tokyo" to a get_weather function. You get standalone scripts, FastAPI servers on ports 8000-8002 with /ask endpoints for auto-routing queries to tools (weather, calculator, contacts), and client scripts that test and log responses with telemetry like tokens/sec and latency. It solves sky-high LLM API costs by proving a $30/month CPU server beats $50k/month cloud bills for 80% of production tasks like intent classification or query parsing.

Why is it gaining traction?

Unlike bloated LLM collections or generic prompt collections on GitHub, this stands out with real benchmarks on RTX 3050 Ti hardware, cost tables showing local wins over GPT-4o-mini, and a tiered architecture demo—start with sub-500ms FunctionGemma routing, escalate to Qwen3 thinkers. Developers grab it for the zero-setup Ollama pulls and pip installs that spin up privacy-safe, offline agents in minutes, mirroring useful github collections like n8n collection github or ansible collection github but laser-focused on tiny LLMs for tool use.

Who should use this?

Backend engineers building agent pipelines tired of API rate limits and vendor lock-in. DevOps teams prototyping high-throughput routers for RAG or MCP skills on edge devices. Indie hackers or startups slashing costs on simple Q&A tools, like replacing cloud calls in IoT or CI/CD with local CPU inference.

Verdict

Solid starter for local tiny LLM experiments—excellent docs, quick starts, and telemetry make it dead simple despite 15 stars and 1.0% credibility score signaling early maturity. Fork and fine-tune for prod; skip if you need battle-tested scale.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.