A Docker-based setup that allows running large language models and speech-to-text on AMD Ryzen AI hardware neural processing units under Linux using FastFlowLM.
How It Works
You hear about an easy way to run smart AI language models super fast using the special AI chip built into your AMD Ryzen computer.
Check that your Linux computer has the right updates, driver, and tools so the AI chip can be accessed safely.
Download the simple package and build it once to create your personal AI environment that talks to the chip.
Choose a clever AI brain like Llama 3.2 and watch it download quickly to your setup for instant use.
Jump into real-time conversations where the AI responds blazingly fast on your hardware.
Create a background service so apps and tools can talk to your AI anytime.
Add speech-to-text to transcribe audio or chat by speaking to the AI.
Delight in generating 60+ words per second with full AI power running locally on your chip, no internet required.
Star Growth
Repurpose is a Pro feature
Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.
Unlock RepurposeSimilar repos coming soon.