sinmb79

Local Korean-first XTTS-v2 engine for one-take shorts narration.

24
1
100% credibility
Found Apr 20, 2026 at 24 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
TypeScript
AI Summary

This repository provides a local tool for generating natural Korean text-to-speech narration using a user-provided voice reference.

How It Works

1
🔍 Discover Korean Voice Tool

You find a handy tool that lets you create smooth Korean voiceovers using your own voice sample, all running on your personal computer.

2
💻 Prepare Your Setup

Follow the simple guide to get everything ready on your Windows computer with one easy preparation step.

3
🎤 Choose Voice Sample

Pick a clear audio clip of Korean speech, at least 10 seconds long without music, to capture the voice you want to use.

4
📝 Write Your Narration Script

Type the Korean text you want spoken into a plain text file, keeping sentences natural for better flow.

5
Generate the Audio

Tell the tool your script and voice sample, add options like speed or length if needed, and it creates a full narration track.

🎉 Enjoy Your Custom Voiceover

Listen to your realistic Korean narration file, ready to use in videos or shorts, with natural phrasing and no service limits.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 24 to 24 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is local-korean-xtts-engine?

This TypeScript CLI tool runs a local Korean-first XTTS-v2 engine to generate one-take narration WAVs for shorts from Korean text files and your own voice reference clips. It fixes multilingual TTS headaches like unnatural Korean pronunciation, token-choppy pacing, and noisy tails by preprocessing text for fluid sentence flow, all offline without SaaS dependencies. Setup via a Windows PowerShell script gets Python 3.11, Torch, and TTS ready; then `npm run synth -- --text-file script.txt --output out.wav --reference ref.wav` spits out polished audio.

Why is it gaining traction?

It stands out as a local GitHub alternative for reproducible voice synthesis—no cloud quotas or policy hassles, just your machine handling XTTS-v2 inference with CUDA/CPU fallback. Korean-specific tweaks like clause-aware line splitting (e.g., `--max-line-length 26`), reference loudness normalization, speed tweaks (`--speed 1.13`), and post-presets for shorts contours deliver pro-sounding output users notice immediately. Multiple references bundle for richer cloning, plus tail cleanup and FFmpeg polish make it a quick local runner for batch narration.

Who should use this?

Korean YouTube shorts creators scripting one-take voiceovers with personal clones. Video editors dodging SaaS edit limits on custom voices. Devs prototyping local Korean TTS pipelines, like in-house narration tools or content automation scripts.

Verdict

Grab it if local Korean XTTS-v2 fits your workflow—strong bilingual docs and CLI make setup straightforward despite 20 stars and 1.0% credibility score signaling early maturity. Test with your refs first; lacks broad testing but nails the niche reliably on Windows/CUDA.

(198 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.