Fantasy-AMAP/fantasy-talking
[ACM MM 2025] FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis
This project helps create realistic talking videos from a still image and an audio file. You provide a photo of a person and an audio recording of what you want them to say, and it generates a video where the person in the photo speaks the words from the audio with natural lip movements and facial expressions. This is ideal for content creators, marketers, or educators looking to animate static images for presentations or social media.
1,622 stars.
Use this if you need to quickly generate a video of a person speaking from just an image and an audio file, with the option to guide their non-verbal communication.
Not ideal if you need to generate full-body animated characters or require highly complex, custom animation beyond talking head videos.
Stars
1,622
Forks
126
Language
Python
License
Apache-2.0
Category
Last pushed
Jan 26, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Fantasy-AMAP/fantasy-talking"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
ModelTC/LightX2V
Light Image Video Generation Inference Framework
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators