MOSTAFA1172m/Image-text-video-I2VGENXL

A PyTorch implementation of a text-image to video diffussion model with a cascaded architecture I2VGEN-XL

22
/ 100
Experimental

This tool helps creative professionals or digital artists transform static images and descriptive text into dynamic video clips. You provide an image and a text prompt (e.g., "Newton smiling and waving"), and it generates a short video animating the scene based on your input. It's ideal for anyone looking to quickly create visual content for social media, presentations, or storytelling.

No commits in the last 6 months.

Use this if you need to animate still images with specific motions or actions described in text, without requiring complex video editing software.

Not ideal if you require precise control over every frame of the video, professional-grade animation quality, or input formats other than a single image and text.

digital-art content-creation video-production visual-storytelling media-design
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

15

Forks

Language

Jupyter Notebook

License

MIT

Last pushed

Jan 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/MOSTAFA1172m/Image-text-video-I2VGENXL"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.