bytedance/LatentSync

Taming Stable Diffusion for Lip Sync!

51
/ 100
Established

This tool helps content creators, marketers, or educators quickly and accurately match spoken audio to a person's mouth movements in a video. You provide an existing video clip of a person speaking and an audio track, and it generates a new video where the person's lips are perfectly synchronized with the provided audio. This is ideal for creating realistic dubbed content, fixing out-of-sync footage, or animating characters to speak specific dialogue.

5,506 stars. No commits in the last 6 months.

Use this if you need to precisely align a person's lip movements in a video with a new or edited audio track, ensuring natural and convincing visual speech.

Not ideal if you need to generate a full human video from scratch, as this tool focuses specifically on the lip-syncing aspect of existing footage.

video-editing dubbing content-creation animation post-production
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 23 / 25

How are scores calculated?

Stars

5,506

Forks

899

Language

Python

License

Apache-2.0

Last pushed

Jun 20, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/bytedance/LatentSync"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.