vbario

A language model that forms persistent memories from conversation and maintains them through sleep. MEMIT weight editing + null-space-constrained maintenance.

27
2
100% credibility
Found Feb 27, 2026 at 16 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

Sleeping LLM is a local AI chatbot that remembers facts from conversations by directly updating its knowledge during chats and reinforcing them through automated rest cycles.

How It Works

1
🖥️ Discover your remembering AI

Find this free tool that lets you chat with an AI that builds real memories from your conversations, just like a friend.

2
💬 Start chatting naturally

Talk about your life, places you live, jobs, favorites — it remembers facts right away without notes or databases.

3
🧠 Memories fill up

After many chats, it notices memories getting crowded and suggests a quick nap to sort them out safely.

4
😴 Let it nap or sleep

Say the word or let it rest — it quietly organizes and strengthens what it learned while you wait.

5
☀️ Wakes up smarter

It comes back refreshed, with all your shared facts locked in forever, even after restarting your computer.

❤️ Lifelong companion grows

Your AI friend now remembers everything from every chat, getting better over time without forgetting.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 16 to 27 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is sleeping-llm?

Sleeping-llm is a Python-based local language model AI that builds persistent memories from conversations by editing weights during "wake" chats and consolidating them via "sleep" cycles. Users chat normally, facts get injected instantly for recall, then sleep commands like /sleep or /nap audit, refresh, and transfer knowledge to stable long-term storage—no databases or retrieval needed. It handles up to 60 facts on 70B models with full recall post-sleep, running on MacBooks or H100s.

Why is it gaining traction?

Unlike RAG systems that stuff context or fetch from external stores, this embeds knowledge directly in weights for zero-latency recall and true generalization across conversations. Developers notice the neuroscience-inspired rhythm: fast wake edits degrade over time, sleep restores without PPL spikes, enabling unbounded capacity as short-term edits dissolve into fused long-term ones. Early benchmarks show 100% chat recall after 2-3 cycles, even as github language model unavailable issues plague Copilot alternatives.

Who should use this?

AI hobbyists running local Llama models who want conversation editing without losing facts between sessions. Researchers testing continual learning on edge hardware like M3 Macs, especially for language model ai experiments on memory retention. Indie devs building personalized copilots that evolve from user interactions, bypassing github language statistics showing Python's dominance in such tools.

Verdict

Worth forking for local LLM tinkerers—solid docs, 5 papers, and repro scripts despite 12 stars and 1.0% credibility score signaling early maturity. Test on 3B for quick wins, but expect tweaks for production as alignment taxes hit larger models.

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.