Osilly

Multimodal deep-research MLLM and benchmark. The first long-horizon multimodal deep-research MLLM, extending the number of reasoning turns to dozens and the number of search-engine interactions to hundreds.

439
44
100% credibility
Found Feb 08, 2026 at 285 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

Research project offering datasets, pre-trained models, and code for training multimodal LLMs specialized in deep visual and textual research tasks, including a new benchmark VDR-Bench.

How It Works

1
🔍 Discover Vision-DeepResearch

You stumble upon this project while looking for tools to help AI handle tough image and text research tasks.

2
📈 Wow, check those results

Videos and charts show models cracking complex visual searches that stump others.

3
📥 Download datasets

Grab free training data and benchmarks from Hugging Face to fuel your experiments.

4
⚙️ Set up your workspace

Follow easy guides to prepare everything for training your own model.

5
🚀 Train the base model

Run simple commands to teach your AI the basics with supervised fine-tuning.

6
🔄 Supercharge with RL

Add reinforcement learning to make your model a deep research expert.

7
📊 Test on the benchmark

Run evaluations to measure how well your model performs on real challenges.

🎉 Your research AI shines

Celebrate as your multimodal model excels at visual and textual deep dives.

Sign up to see the full architecture

6 more

Sign Up Free

Star Growth

See how this repo grew from 285 to 439 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is Vision-DeepResearch?

Vision-DeepResearch is a Python-based multimodal LLM framework and benchmark for long-horizon deep-research tasks, the first to extend reasoning turns to dozens and search-engine interactions to hundreds. It lets you train models like the 8B and 30B-A3B variants using SFT and RL on vision-text datasets from Hugging Face, tackling complex queries that mix images, web searches, and iterative analysis. Developers get pretrained weights, toy SFT/RL datasets, and VDR-Bench for evaluation across multimodal datasets github-style workflows.

Why is it gaining traction?

It stands out as the first github multimodal llm pushing multimodal deep research boundaries, crushing baselines on VDR-Bench with 50.5% average score for the 8B model versus 40.1% agentic rivals. Hooks include async RL pipelines for hundreds of interactions, demo videos showing real-world research beats, and easy vLLM serving for inference—perfect for multimodal rag pipeline github experiments without custom agents.

Who should use this?

AI researchers fine-tuning multimodal deep researchers for visual question-answering or web analysis. Teams building multimodal github copilot-like tools or multimodal rag langchain github integrations that need dozens of reasoning steps. Benchmark enthusiasts testing phi4 multimodal github models on deepresearch scenarios.

Verdict

Promising for multimodal universe explorers, with solid HF datasets and scripts—but at 1.0% credibility score and 315 stars, it's early-stage with incomplete docs. Try the 8B demo if long-horizon vision research fits; otherwise, wait for full 30B release.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.