beanie00 / self-distillation-analysis
PublicCodebase for the work “Why Does Self-Distillation (Sometimes) Degrade the Reasoning Capability of LLMs?”
Research codebase reproducing why self-distillation degrades LLM reasoning by suppressing uncertainty expression, with scripts for analysis, data prep, model training, and evaluation on math tasks.
How It Works
You stumble upon research showing why copying perfect answers sometimes makes AI worse at math puzzles.
Read the clear instructions to set up your tools and grab sample math problems.
Run quick checks to watch how confident hints make AI lose its hesitant, exploratory style.
Dive deeper into reports on uncertainty words fading away.
Launch experiments blending hints with original thinking.
Compare before-and-after results on tough unseen problems.
Your AI now tackles harder math with balanced confidence and exploration.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.