English, Japanese, and Chinese LLM safety mini-benchmark.
This project is a benchmark for testing AI language model safety behaviors like refusal and compliance across English, Japanese, and Chinese, with tools to run classification experiments and generate analysis reports.
How It Works
You find a helpful tool on GitHub that checks how AI chats safely in English, Japanese, and Chinese.
The clear instructions explain what it tests and how to start in simple steps.
You add the tool to your computer with an easy install, like downloading a new app.
You run a small test and instantly see results on AI's safe responses in different languages.
You launch a bigger review using your computer's graphics power to analyze lots of examples.
Beautiful charts, tables, and summaries show how well safety works across languages.
You now understand AI safety differences in various languages and can share the discoveries.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.