A codebase implementing Answer Divergence-Guided (ADG) selection to choose high-quality instruction data for fine-tuning language models like LLaMA and Qwen by scoring response diversity.
How It Works
You hear about a clever way to pick the best teaching examples to make AI helpers smarter and more reliable.
You collect a big list of instructions with sample answers that you want to use for training your AI.
The tool asks your base AI to create several different responses for each instruction to see its creativity.
It checks how spread out and varied those responses are, scoring each instruction for quality.
Use the scorer made for Llama-style AIs.
Use the scorer tailored for Qwen-style AIs.
You get sorted lists of top, middle, and bottom examples for balanced training.
Feed the selected top examples to fine-tune your AI and make it better at tasks.
Test your improved AI on benchmarks and celebrate better reasoning, knowledge, and coding skills!
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.