sseanliu

Real-time AI assistant for Meta Ray-Ban smart glasses -- voice + vision + agentic actions via Gemini Live and OpenClaw

1,408
272
100% credibility
Found Feb 07, 2026 at 201 stars 7x -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
AI Summary

VisionClaw is an iOS app that provides real-time voice conversations with visual understanding through Meta Ray-Ban smart glasses or an iPhone camera, optionally enabling actions like messaging or searches via connected tools.

How It Works

1
πŸ” Discover VisionClaw

You come across VisionClaw, a fun app that turns your smart glasses or phone into a talking AI buddy that sees what you see.

2
πŸ“± Prepare the app

Open the app on your computer, connect a free AI service, and get it ready to run on your iPhone.

3
▢️ Test with your phone

Tap start on your iPhone to use its camera and mic for chatting with the AI.

4
πŸ—£οΈ Chat with AI vision

Speak naturally like 'What am I looking at?' and hear the AI describe the world right through your screen.

5
Pick your power-up
πŸ‘“
Add smart glasses

Connect your Ray-Ban glasses so the AI sees live from what you're wearing.

πŸ› οΈ
Add action tools

Link a helper on your computer to let the AI send messages, search, or manage lists.

βœ…
Keep it simple

Enjoy voice and vision chats just as they are.

πŸŽ‰ Your AI companion lives!

Now talk to your glasses or phone anytime – it sees, responds, and helps with daily life effortlessly.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 201 to 1,408 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is VisionClaw?

VisionClaw turns Meta Ray-Ban smart glasses into a real-time AI assistant that sees through the camera, hears your voice, and executes actions like adding shopping items or sending messages. Built in Swift for iOS, it streams low-fps video and bidirectional audio to Google's Gemini Live API over WebSocket, with optional OpenClaw integration for 50+ app-connected tools. Test it instantly on your iPhone camera without glasses, using a free key from Google AI Studio.

Why is it gaining traction?

It delivers seamless real-time assistant AIβ€”voice queries like "What am I looking at?" get instant visual descriptions, while agentic tasks route through OpenClaw for web searches or messaging. Developers love the quick start: clone, add API key, run on device, no complex setup. Standing out from basic real-time transcription GitHub projects, it combines vision, voice, and actions in one pipeline.

Who should use this?

iOS devs prototyping wearable AI, hardware hackers with Ray-Bans, or anyone building real-time assistants like medical instructed aids for glaucoma and diabetic patients, real-time poker or interview assistants, meeting note-takers, or sales bots. Ideal for real-time detection projects or power monitoring dashboards needing voice-vision integration.

Verdict

Grab it if you're in wearables or real-time AIβ€”607 stars show buzz, and docs make setup straightforward despite 1.0% credibility score signaling early maturity. Solid for proofs-of-concept, but watch for production hardening.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.