qlabs-eng

qlabs-eng / slowrun

Public

100M tokens, no time limit, best val loss wins!

115
21
100% credibility
Found Feb 27, 2026 at 103 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

A competition and tools for training language models on a fixed small dataset of 100M tokens to minimize validation loss with unlimited or limited compute.

How It Works

1
🔍 Discover the Slowrun Challenge

You find this fun contest where people train AI to predict stories using a fixed set of writings and unlimited computer time to get the best score.

2
📖 Read the Guide and Leaderboards

Learn the simple rules, see top scores, and decide if you'll try the no-limits path or the quick one-hour challenge.

3
📦 Get the Stories Ready

Use the easy prep tool to download and chop up the writings into bite-sized pieces for your AI to learn from.

4
🚀 Launch the Training

Start the learning session on your powerful computers and let the AI practice predicting words over and over.

5
📈 Watch It Improve

Thrillingly see the error numbers drop lower and lower as your AI masters the stories better than before.

6
📊 Check Your Final Score

Once done, look at your best prediction accuracy and compare it to everyone else's on the leaderboard.

🎉 Share Your Breakthrough

Celebrate by posting your winning score for others to try and beat, joining the community fun.

Sign up to see the full architecture

5 more

Sign Up Free

Star Growth

See how this repo grew from 103 to 115 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is slowrun?

Slowrun is a Python benchmark for training transformer language models on a fixed 100M tokens from FineWeb, with no time limit—lowest validation loss wins. It flips the script on speedruns like modded-nanogpt by prioritizing compute-heavy techniques over wall-clock speed, letting you hammer 100M tokens for maximum learning. Run python prepare_data.py to tokenize the data, then torchrun train.py on H100s for the 2.7B baseline hitting 3.402 val loss; submit PRs to leaderboards for unlimited or 1-hour limited tracks.

Why is it gaining traction?

Unlike data-rich speedruns, slowrun unlocks heavy regularization, exotic optimizers like Muon, and massive overparameterization on github 100m datasets, potentially yielding GPT-3-style leaps. Karpathy's endorsement adds cred, and the simple PR-based leaderboard hooks competitive devs chasing best val loss without infinite data worries. Early traction from its contrast to prithvi 100m github or tahoe 100m github experiments.

Who should use this?

ML engineers with H100 clusters testing optimizers or scaling laws in data-limited regimes. Researchers benchmarking regularization on 100M token caps, or teams iterating slowrun speedrun hybrids for production LM fine-tunes. Avoid if you're on consumer GPUs—needs Hopper for Flash Attention.

Verdict

Grab it if you have the hardware to beat the baseline; 84 stars and 1.0% credibility score signal an early, barebones benchmark with solid docs but no tests or broad validation yet. Fork and PR your improvements—low barrier for real contributions.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.