adam-s / testing-claude-agent
PublicDeterministic benchmarking of .claude/ instruction sets for Claude Code token efficiency
This repository benchmarks various instruction configurations for Claude AI coding agents by measuring token consumption required to solve standardized programming challenges with passing automated tests.
How It Works
You stumble upon this GitHub page while reading discussions about making AI helpers more efficient at coding tasks.
You explore the tests comparing different ways to guide an AI through coding puzzles, seeing clear tables of winners and costs.
You get excited spotting which simple instructions let the AI solve challenges using the least effort and money.
You create a few coding challenges with ready checks to see if solutions work perfectly.
You launch tests for each instruction style, watching the AI tackle your puzzles one by one.
You check easy charts and summaries showing speeds, costs, and which style shines brightest.
Now you know the smartest way to instruct your AI helper for future coding adventures, saving time and cash.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.