ShivamDuggal4/UNITE-tokenization-generation
Single-stage End-to-End Training for Tokenization and Generation
This project helps machine learning researchers efficiently train generative models that create high-fidelity images or molecular structures. It takes raw image datasets (like ImageNet) or molecule data as input and produces a trained model capable of both converting images into a compressed 'latent' representation (tokenization) and generating new, realistic images or molecules from text descriptions or noise. This is ideal for researchers developing new latent diffusion models or exploring efficient training methods for generative AI.
Use this if you are a researcher focused on developing or improving large-scale image and molecule generation models, and you want to streamline the training process by unifying tokenization and generation.
Not ideal if you are a practitioner looking for an out-of-the-box tool to generate images without deep involvement in model architecture and training.
Stars
62
Forks
1
Language
Python
License
—
Category
Last pushed
Mar 24, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/ShivamDuggal4/UNITE-tokenization-generation"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
NVlabs/Sana
SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer
FoundationVision/VAR
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈]...
nerdyrodent/VQGAN-CLIP
Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.
huggingface/finetrainers
Scalable and memory-optimized training of diffusion models
AssemblyAI-Community/MinImagen
MinImagen: A minimal implementation of the Imagen text-to-image model