MOSS-TTSD and MOSS-TTS
MOSS-TTSD is a specialized extension of MOSS-TTS that adds dialogue-specific capabilities (long-context modeling, multi-speaker synthesis) on top of the core TTS functionality, making them ecosystem siblings where MOSS-TTSD builds upon MOSS-TTS for conversational applications.
About MOSS-TTSD
OpenMOSS/MOSS-TTSD
MOSS-TTSD is a spoken dialogue generation model designed for expressive multi-speaker synthesis. It features long-context modeling, flexible speaker control, and multilingual support, while enabling zero-shot voice cloning from short audio references.
This project helps content creators transform dialogue scripts into dynamic, expressive spoken conversations with multiple distinct speakers. You provide a script and short audio references for each speaker, and it generates natural-sounding, long-form spoken dialogue up to 60 minutes. It's ideal for producers of podcasts, audiobooks, commentary, and dubbed content.
About MOSS-TTS
OpenMOSS/MOSS-TTS
MOSS‑TTS Family is an open‑source speech and sound generation model family from MOSI.AI and the OpenMOSS team. It is designed for high‑fidelity, high‑expressiveness, and complex real‑world scenarios, covering stable long‑form speech, multi‑speaker dialogue, voice/character design, environmental sound effects, and real‑time streaming TTS.
The MOSS-TTS Family helps you create incredibly realistic and expressive speech and sound effects from text. You provide text and receive high-quality audio that sounds like a real person, handles multiple speakers, and can even generate unique voices or environmental sounds. This is perfect for content creators, game developers, virtual assistant designers, and anyone needing advanced audio generation.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work