MOSTAFA1172m/Image-text-video-I2VGENXL
A PyTorch implementation of a text-image to video diffussion model with a cascaded architecture I2VGEN-XL
This tool helps creative professionals or digital artists transform static images and descriptive text into dynamic video clips. You provide an image and a text prompt (e.g., "Newton smiling and waving"), and it generates a short video animating the scene based on your input. It's ideal for anyone looking to quickly create visual content for social media, presentations, or storytelling.
No commits in the last 6 months.
Use this if you need to animate still images with specific motions or actions described in text, without requiring complex video editing software.
Not ideal if you require precise control over every frame of the video, professional-grade animation quality, or input formats other than a single image and text.
Stars
15
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Jan 08, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/MOSTAFA1172m/Image-text-video-I2VGENXL"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
ModelTC/LightX2V
Light Image Video Generation Inference Framework
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators