DestinyLinker

A benchmark for evaluating LLMs on Chinese traditional fortune telling — Bazi (八字) and Ziwei Doushu (紫微斗数).

30
1
100% credibility
Found Apr 28, 2026 at 30 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

MingLi-Bench is an evaluation tool and dataset of 160 multiple-choice questions from Chinese fortune telling competitions, used to score how accurately large language models predict life events via Bazi and Ziwei Doushu.

How It Works

1
🔍 Discover the fortune telling test

You hear about this fun benchmark that checks how well smart AI helpers predict life events using ancient Chinese wisdom like birth charts.

2
📥 Bring it home

Download the simple tool and get it ready on your computer in just a few minutes.

3
🔗 Link your AI friends

Connect popular AI services you like, so they can join in answering the fortune questions.

4
🚀 Launch a test run

Choose a year or topic like career or love, turn on thinking steps if you want, and let the AIs tackle real fortune telling puzzles from competitions.

5
📊 Check the scores

Watch a progress bar as results pour in, showing right and wrong answers with details on speed and categories.

🎉 Crown the champion

Celebrate seeing which AI is the best at Chinese fortune telling, with summaries and files to share your discoveries.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 30 to 30 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is MingLi-Bench?

MingLi-Bench is a Python CLI benchmark evaluating large language models on Chinese traditional fortune telling—Bazi and Ziwei Doushu—with 160 multiple-choice questions from global competitions (2022-2025). It tests LLMs on life predictions like career, marriage, and wealth, using exact-match scoring and options like Chain-of-Thought prompting or pre-computed astrological charts to isolate reasoning from chart generation. Run `python -m mingli_bench.cli --model gpt-4o --cot --astro` for instant results with JSON summaries and raw responses.

Why is it gaining traction?

This stands out in the bazi benchmark space by supporting OpenRouter for one-key access to dozens of models, plus native APIs for DeepSeek, Doubao, and others—filter by year or category like `--categories 事业 婚姻`. Developers love the concurrency (`--max-workers 8`), option shuffling to kill position bias, and artifacts like per-question files for debugging LLM failures. It's a quick way to probe cultural reasoning gaps without building your own eval suite.

Who should use this?

LLM researchers benchmarking models on non-Western domains, Chinese AI teams evaluating local providers like Doubao on traditional knowledge, or anyone auditing frontier models for fortune-telling accuracy. Ideal for ablation studies on CoT or astrology injection in prompts.

Verdict

Grab it if you're into niche LLM evals—solid CLI and docs make it dead simple, despite 30 stars and 1.0% credibility signaling early days. MIT-licensed and production-ready for quick tests, but watch for dataset expansions.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.