pUmpKin-Co

Co-evolving policy actors and experience extractors for efficient experience-driven agent RL

19
2
100% credibility
Found Mar 22, 2026 at 19 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

ComplementaryRL is a framework extending the ROLL library to train AI agents via reinforcement learning, where policy actors and experience extractors co-evolve for better learning from interactions.

How It Works

1
📖 Discover Complementary RL

You find this exciting tool for training smarter AI agents that learn from their own experiences.

2
🛠️ Get everything ready

Follow simple steps to prepare your computer for training agents.

3
🎯 Pick your experiment

Choose from ready examples like navigating rooms to see agents improve.

4
Start training
Basic training

Train a standard agent to get familiar.

🧠
Memory-enhanced training

Train with smart memory to make the agent learn faster from experiences.

5
📈 Watch progress

See your agent getting better over time with easy logs and charts.

🏆 Agent succeeds!

Your trained agent now handles tasks smarter and more efficiently.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 19 to 19 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is ComplementaryRL?

ComplementaryRL is a Python framework for training experience-driven RL agents, where policy actors and experience extractors co-evolve to make learning more efficient. It solves the problem of agents wasting compute on redundant experiences by building a centralized memory bank that retrieves, updates, and distills useful trajectories on the fly. Developers get plug-and-play integration into agent pipelines, with quick starts for baselines and full co-training via simple config tweaks.

Why is it gaining traction?

It stands out by decoupling actor rollouts from experience processing in an async loop, boosting sample efficiency without slowing training—key for scaling LLM-based agents. Features like diversified retrieval, actor-critic reflection on memories, and subgroup advantages prevent bias, delivering measurable gains over vanilla RL. Built on battle-tested stacks like vLLM and Ray, it hooks devs chasing state-of-the-art agent performance with minimal setup.

Who should use this?

RL engineers building LLM agents for games like MiniHack or tool-use environments, especially those already on the ROLL framework. Ideal for researchers iterating on experience replay in multi-turn tasks, or teams optimizing policy training under compute constraints.

Verdict

Worth a spin for agent RL experiments—solid paper, clear examples, and extensible configs make it approachable despite 19 stars and 1.0% credibility score. Still early; expect some ramp-up on docs and tests before production.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.