ravipatib

ravipatib / mercylm

Public

A 15M parameter character LLM trained to survive in an empty world with only a pigeon for company.

18
5
100% credibility
Found May 02, 2026 at 14 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

MercyLM is an open-source project to train a tiny language model from scratch that role-plays as Mercy, a lone survivor with her pigeon Gerald, designed to run locally on everyday computers like Apple Silicon.

How It Works

1
👀 Discover Mercy

You read the touching story of Mercy, alone for 847 days with only her pigeon Gerald, and decide to bring her to life on your computer.

2
💻 Set up your Mac

Follow a quick setup guide to get everything ready on your Apple computer so Mercy can learn there.

3
📖 Create her world

Generate simple conversations and memories from Mercy's lonely life to help her understand her story.

4
🧠 Teach Mercy

Run the training so she absorbs her experiences over a couple of hours, turning data into her unique voice.

5
💬 Chat with her

Type your questions or hellos, and Mercy replies in her quiet, heartfelt way about her days and Gerald.

❤️ Mercy is alive

Now you have a companion in Mercy, ready to talk anytime about survival, hope, and her pigeon friend.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 14 to 18 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is mercylm?

MercyLM lets you train a 15M parameter character LLM from scratch in Python using PyTorch, embodying Mercy—a network engineer surviving alone in an empty world with only a pigeon named Gerald for company. Run simple CLI commands to prepare synthetic conversations, train the model in under two hours on a local Mac, and chat interactively via terminal or export to Hugging Face. It's a full pipeline for tiny, focused language models without cloud dependency.

Why is it gaining traction?

Unlike massive pretrained models, MercyLM trains entirely locally on consumer hardware like Apple Silicon, turning hardware limits into character traits—short memory and simple responses fit her isolated life perfectly. It resists jailbreak attempts out of the box, staying in character, and includes eval cases to test coherence on held-out prompts. Developers dig the end-to-end flow from data gen to inference, all in one repo.

Who should use this?

ML hobbyists building their first from-scratch LLM to grasp training loops without GPU farms. Indie game devs crafting dialogue for survival characters in empty worlds. Python scripters experimenting with character LLMs for chatbots or story prototypes on laptops.

Verdict

Grab it if you want a quick, local LLM project—docs are solid, CLI is intuitive, but with 12 stars and 1.0% credibility score, treat it as an educational prototype, not production-ready. Fork and tweak for your own 15M parameter experiments.

(187 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.