mega-edo / mega-security-leaderboard
PublicLeaderboard Comparing LLM Agent Security on System Prompt Leakage and Attack Probes
A leaderboard benchmarking how effectively various AI language models protect their core instructions from attacks such as jailbreaks, data leaks, and injections, both before and after prompt optimization.
How It Works
You find this webpage that compares how well popular AI chatbots resist sneaky tricks trying to reveal or override their hidden instructions.
You quickly see the main finding: simple tweaks to instructions make even smaller, cheaper AIs safer than big ones out of the box.
You notice small AIs with tuned instructions topping the charts, beating expensive giants in blocking attacks.
You browse tables showing scores for different AI makers, model sizes, and attack types like fake personas or data grabs.
You understand that custom tuning your AI's instructions beats just picking a powerful model.
Now you know which setups work best and how to test your own AI assistant for real-world protection.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.