hyp1231

hyp1231 / Latte

Public

Implementations for "Expressiveness Limits of Autoregressive Semantic ID Generation in Generative Recommendation"

19
1
100% credibility
Found May 15, 2026 at 19 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
Python
AI Summary

This repository implements Latte, a research model that enhances generative product recommendations by inserting latent tokens before semantic item identifiers, tested on Amazon review datasets.

How It Works

1
📖 Discover Latte

You stumble upon this project through a research paper promising smarter product suggestions by unlocking hidden patterns in reviews.

2
💻 Set up your space

Grab the files and prepare your computer with a quick setup so you can start playing around right away.

3
Pick a product world
🎮
Video Games

Dive into gaming reviews for personalized game suggestions.

🔬
Science Tools

Explore industrial gear recommendations from expert feedback.

🎸
Music Gear

Tune into instruments and gear picks from music lovers.

4
🚀 Launch the learning

Hit start on training and watch your recommender brain grow smarter from thousands of customer stories.

5
📈 Review the magic

Check the scores and charts showing how much better your suggestions hit the mark compared to before.

🎉 Better picks unlocked

Celebrate as your new recommender uncovers subtle patterns for spot-on product matches that delight users.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 19 to 19 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is Latte?

Latte implements a T5-based generative recommendation model that boosts expressiveness in autoregressive semantic ID generation, tackling limitations where standard GR models fail to capture simple collaborative filtering patterns. You train it on Amazon Reviews 2023 categories like Industrial_and_Scientific or Video_Games, generating conflict-free item IDs via vector quantization (Faiss OPQ, RQ K-means, or RQ-VAE) from sentence embeddings. Python setup with uv sync and one-liner training via train_latte.sh delivers 3.45% NDCG@10 gains over baselines—ideal for dl paper implementations github like this arXiv work.

Why is it gaining traction?

Unlike pure semantic ID baselines (PSID here), Latte prepends learnable latent tokens to sidestep decoding bottlenecks, with agg_max/sum for multi-path aggregation during beam search (num_beams=50-500). Quickstart scripts handle caching, inference, and WandB/TensorBoard logging, letting you swap VQ methods or resume checkpoints seamlessly. Developers grab it for reproducible recsys experiments without rebuilding tokenizers from scratch.

Who should use this?

Recsys researchers replicating generative retrieval papers, ML engineers benchmarking semantic compression on text-rich catalogs like e-commerce reviews, or GenAI teams exploring agentic recommendation limits. Perfect if you're tuning T5 decoders on leave-one-out splits for categories like Musical_Instruments.

Verdict

Solid for academic repros—run main.py --model=Latte --vq_method=rqkmeans and evaluate instantly—but 19 stars and 1.0% credibility signal early-stage maturity; docs are README-focused with no tests. Worth forking for latte review github if you're in dl paper implementations github, else wait for more adoption.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.