RightNow-AI / autokernel
PublicAutoresearch for GPU kernels. Give it any PyTorch model, go to sleep, wake up to optimized Triton kernels.
AutoKernel is a tool that profiles PyTorch models on GPUs to identify bottleneck operations, extracts them as editable Triton kernels, and enables AI agents to autonomously optimize them for speedups via an automated edit-test-revert loop.
How It Works
You hear about a smart tool that finds slow parts in your AI model and uses AI to make them faster overnight.
Download the tool and prepare sample data with a few simple steps on your computer with a good graphics card.
Run your AI model through the profiler to see exactly which math operations are taking the most time on your graphics card.
The tool grabs the slowest operations and turns them into simple building blocks ready for improvement.
Give the AI clear instructions, point it at one block, and let it experiment and improve automatically while you sleep.
Put the improved blocks back into your model, run a check to confirm it's correct, and measure the overall speedup.
Your model now runs quicker end-to-end, saving time and energy on every use, with graphs showing your progress.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.