HKUST-C4G

The official code of "Beyond VLM-Based Rewards: Diffusion-Native Latent Reward Modeling"

46
1
100% credibility
Found Feb 24, 2026 at 41 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

A toolkit for scoring AI-generated images on how closely they match text prompts using a diffusion-based reward model.

How It Works

1
🔍 Discover the scorer

You find this handy tool on GitHub that helps rate AI-generated pictures by how well they match your ideas.

2
📦 Set it up easily

You create a simple workspace on your computer following the quick guide, and everything installs smoothly.

3
🔗 Link your image maker

You connect it to your favorite AI image generator so it understands the pictures it creates.

4
Grab the smart scorer

With one command, you load the ready-trained brain that knows good pictures from great ones.

5
Pick your pictures
🎨
Create new ones

Tell it what to imagine and generate pictures on the spot to score instantly.

🖼️
Use your files

Load pictures from your folder to see how they rate against your description.

6
📝 Add your description

Type a simple sentence about what the perfect picture should show.

Get your scores

You receive clear numbers showing which pictures best capture your vision, ready to pick favorites!

Sign up to see the full architecture

5 more

Sign Up Free

Star Growth

See how this repo grew from 41 to 46 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is diffusion-rm?

diffusion-rm is a Python toolkit for building diffusion-native reward models that score image latents directly during generation, skipping costly VLM evaluations for cheaper alignment. Plug it into Stable Diffusion 3.5 pipelines to train on preference pairs or run inference on generated latents or local images via simple API calls like scorer.reward(text_conds, latents, u=0.4). It delivers normalized scores out-of-the-box, with pretrained models on Hugging Face.

Why is it gaining traction?

It matches VLM reward accuracy at lower compute—key for scaling diffusion alignment beyond 20 github prototypes—while offering plug-and-play inference that integrates with diffusers pipelines. Clear examples for scoring on-the-fly latents or disk images lower the barrier versus custom reward hacks. Research from HKUST shows competitive Thurstone loss performance, hooking devs tired of VLM latency.

Who should use this?

Diffusion fine-tuners aligning SD3 outputs to human prefs without VLM overhead. RLHF teams prototyping reward signals for image gen pipelines. Researchers beyond github nanogpt baselines needing diffusion RM for custom backbones like Flux.

Verdict

Grab it for inference today—pretrained model works seamlessly—but skip full training until dataset and eval code drop per the roadmap. 41 stars and 1.0% credibility score signal early-stage research code; solid docs make it usable now for diffusion RM experiments.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.