lechmazur / persuasion
PublicLLM Persuasion Benchmark tests whether one language model can change another model’s stated position over the course of a multi-turn conversation. It runs round-robin persuasion dialogues on contested propositions and measures both persuasive effectiveness and target resistance from stance shifts recorded before and after each exchange.
This repository hosts a benchmark evaluating large language models' persuasion effectiveness in multi-turn conversations, complete with leaderboards, charts, model profiles, and debate transcripts.
How It Works
You find this GitHub page sharing rankings of AI models' abilities to persuade each other in debates.
You check the top lists to see which AIs are best at changing minds and which resist the most.
Standout models like GPT and Claude shine as powerful persuaders in the colorful charts.
You look at graphs showing head-to-head battles and overall offense versus defense strengths.
You dive into profiles explaining how each AI argues or defends in real conversations.
You read fun transcripts of AIs debating topics like city bans or animal reintroduction.
You now understand which AIs win arguments and why, ready to pick the best for tough talks.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.