HuangShengZeBlueSky

基于 Karpathy AutoResearch 思想,让 AI Agent 自主迭代优化 MLP 模型

17
0
100% credibility
Found Mar 13, 2026 at 17 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

Educational project enabling AI agents to autonomously iterate and optimize multi-layer perceptron models for MNIST handwritten digit classification.

How It Works

1
🔍 Discover MLP AutoResearch

You stumble upon this exciting project that teaches AI to automatically improve simple models for recognizing handwritten digits.

2
💻 Prepare your computer

Follow easy setup steps to get everything ready on your regular computer—no special hardware needed.

3
▶️ Try the starting model

Run the first quick training session and check how well it identifies digits out of the box.

4
🤖 Hand over to AI assistant

Share the simple guide with your favorite AI like Claude or GPT, and let it start making smart changes.

5
🔄 See AI experiment and improve

Watch as the AI tests new ideas, trains models, and keeps only the better versions in a repeating cycle.

🎉 Achieve great digit recognition

Celebrate as your model hits top accuracy scores on recognizing numbers, all while you sit back and relax.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 17 to 17 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is MLP_AutoResearch?

MLP_AutoResearch ports Karpathy's AutoResearch idea from his github projects like nanogpt and llm-council to simple MLPs in Python with PyTorch. It lets AI agents like Claude or GPT autonomously tweak model architectures, hyperparameters, and optimizers on MNIST digit classification, training fixed 20-epoch runs and keeping only improvements based on test accuracy. Developers get a hands-off loop: set rules via a prompt, run baseline experiments with a single command, and watch agents iterate via git commits.

Why is it gaining traction?

Unlike manual tuning in Karpathy's github llm or makemore repos, this runs on CPU without GPUs, making agent-driven optimization accessible beyond H100 clusters. The hook is pure experimentation—agents explore residuals, activations like ReLU or GELU, batchnorm, and schedulers in a greedy loop, logging results to TSV for easy tracking. It's a lightweight bridge from Karpathy's rnn effectiveness studies to MLP ablation.

Who should use this?

ML students replicating deep MLP papers on MNIST, researchers prototyping agent workflows before scaling to Karpathy-style llm-c or nanochat setups, or Python devs curious about autoresearch for hyperparameter sweeps. Ideal for those tired of scripting grid searches in notebooks.

Verdict

Fun teaching project with solid docs and easy uv/pip setup, but at 17 stars and 1.0% credibility score, it's immature—results are TBD, no tests. Try for learning Karpathy agent patterns, skip for production.

(178 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.