lucas-maes / le-wm
PublicOfficial code base for LeWorldModel: Stable End-to-End Joint-Embedding Predictive Architecture from Pixels
LeWorldModel is a research codebase for training and evaluating compact AI world models that predict future states from raw pixel observations in robotics tasks.
How It Works
You stumble upon this cool research project that teaches AI to predict how objects move just by watching pixel videos, perfect for robot brains.
You set up a simple virtual space on your computer to get everything ready for training the AI.
You grab ready-made packs of robot action videos from a shared online folder and place them in your storage spot.
You start the training process, feeding it videos so it learns to imagine future movements – it hums along for a few hours on your computer's graphics power.
You load your trained AI into test environments like pushing objects or navigating rooms to see how well it plans actions.
Your AI now reliably forecasts movements, detects weird events, and controls robots better and faster than before, ready for real-world experiments.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.