tdemin16 / proactivebench
PublicOfficial repository of "ProactiveBench: Benchmarking Proactiveness in Multimodal Large Language Models"
ProactiveBench is a benchmark with tools and datasets to evaluate if image-understanding AI models proactively request simple user actions to resolve unclear visuals like occlusions or distortions.
How It Works
You stumble upon this benchmark while exploring ways to test if picture-seeing AI helpers know when to ask for better views on tricky images.
You easily add the benchmark kit to your computer with a quick install, getting everything ready to go.
You download sets of challenging images—like blocked objects, blurry shots, or rough sketches—from a shared online collection.
You feed the tough images to your AI model and watch if it smartly suggests fixes like moving blocks or enhancing clarity before guessing.
You get simple scores showing the AI's accuracy and how often it proactively asks for help to see better.
Now you clearly see how proactive your AI is at handling real-life visual puzzles, ready to improve it further!
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.