kyegomez/LUMIERE
Implementation of the text to video model LUMIERE from the paper: "A Space-Time Diffusion Model for Video Generation" by Google Research
This project helps researchers and developers explore and implement advanced video generation techniques. It takes text descriptions as input and aims to produce corresponding video content. The primary users are individuals working on AI research and development, particularly those interested in video synthesis and diffusion models.
No commits in the last 6 months. Available on PyPI.
Use this if you are an AI researcher or developer looking to experiment with and build upon a specific component of a text-to-video diffusion model.
Not ideal if you are looking for an out-of-the-box application to generate videos without coding or deep technical understanding.
Stars
52
Forks
5
Language
Python
License
MIT
Category
Last pushed
Jan 27, 2025
Commits (30d)
0
Dependencies
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/kyegomez/LUMIERE"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
NVlabs/Sana
SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer
FoundationVision/VAR
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈]...
nerdyrodent/VQGAN-CLIP
Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.
huggingface/finetrainers
Scalable and memory-optimized training of diffusion models
AssemblyAI-Community/MinImagen
MinImagen: A minimal implementation of the Imagen text-to-image model