Testing a few local models on ability to understand a research paper and accompanying code.
This repository shares a detailed evaluation of local AI models' performance in mapping computational simulation code to descriptions in a corresponding research paper, including outputs, corrections, and key findings.
How It Works
You stumble upon this collection while searching for insights on how smart computer brains match computer programs to science papers.
You scan the main page to learn about tests on different computer thinkers linking code recipes to research stories.
You get excited seeing which local thinkers like Qwen shine brightest at spotting the right connections.
You open files to see raw attempts, fixes, and perfect matches between code parts and paper ideas.
You review what went well and what tripped them up, like missing tiny details.
You walk away knowing local computer brains are catching up fast, making science checks easier for everyone.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.