OpenMOSS/MOVA
MOVA: Towards Scalable and Synchronized Video–Audio Generation
This project helps content creators, marketers, or educators quickly generate realistic videos with perfectly synchronized speech and relevant sound effects. You provide a text description and an optional reference image, and it outputs a high-fidelity video clip with synchronized audio. This tool is ideal for anyone needing to create compelling video content efficiently without complex editing.
833 stars. Actively maintained with 2 commits in the last 30 days.
Use this if you need to create engaging video content, especially with spoken dialogue, where precise lip-syncing and contextual sound effects are crucial for realism.
Not ideal if you need to edit existing video footage or require highly artistic, stylized video generation beyond realistic scenarios.
Stars
833
Forks
53
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 14, 2026
Commits (30d)
2
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/OpenMOSS/MOVA"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
ModelTC/LightX2V
Light Image Video Generation Inference Framework
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators