Multi-Tacotron-Voice-Cloning and Cross-Lingual-Voice-Cloning

Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 24/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 20/25
Stars: 397
Forks: 91
Downloads:
Commits (30d): 0
Language: Python
License:
Stars: 360
Forks: 56
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: BSD-3-Clause
Archived Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About Multi-Tacotron-Voice-Cloning

vlomme/Multi-Tacotron-Voice-Cloning

Phoneme multilingual(Russian-English) voice cloning based on

Implements a four-stage pipeline combining speaker verification (GE2E encoder), phonemic text-to-speech synthesis (Tacotron 2), and neural vocoding (WaveRNN) to enable few-shot voice cloning from seconds of audio. Uses phoneme-based representation with language-specific dictionaries to support both Russian and English in a unified model. Provides pretrained weights and curated multilingual datasets, with training extensible to additional languages via the phoneme dictionary approach.

About Cross-Lingual-Voice-Cloning

deterministic-algorithms-lab/Cross-Lingual-Voice-Cloning

Tacotron 2 - PyTorch implementation with faster-than-realtime inference modified to enable cross lingual voice cloning.

This tool helps content creators and educators generate speech in a new language using an existing speaker's voice. You provide audio files of a person speaking along with corresponding text, and the system learns to clone their voice. The output is a synthesized audio file of that speaker's voice speaking a different language from the input, making it useful for creating consistent voiceovers or narrated content across multiple languages.

voice-over-production multilingual-content e-learning-localization audio-post-production synthetic-media

Scores updated daily from GitHub, PyPI, and npm data. How scores work