pytorch-vae and pytorch-vq-vae

These are complementary implementations exploring different VAE architectures—the basic VAE provides a foundation for understanding variational inference, while VQ-VAE extends it with vector quantization to learn discrete latent representations, making them useful for different use cases rather than interchangeable alternatives.

pytorch-vae
50
Established
pytorch-vq-vae
48
Emerging
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 24/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 22/25
Stars: 432
Forks: 107
Downloads:
Commits (30d): 0
Language: Python
License: BSD-3-Clause
Stars: 602
Forks: 102
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: MIT
Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About pytorch-vae

ethanluoyc/pytorch-vae

A Variational Autoencoder (VAE) implemented in PyTorch

This is a foundational building block for machine learning engineers and researchers working with deep learning models. It takes in complex data, like images or text, and learns a compressed, meaningful representation of that data. This compressed representation can then be used for generating new, similar data, or for tasks like anomaly detection.

deep-learning generative-modeling data-compression representation-learning anomaly-detection

About pytorch-vq-vae

zalandoresearch/pytorch-vq-vae

PyTorch implementation of VQ-VAE by Aäron van den Oord et al.

This is a PyTorch implementation of VQ-VAE. It is a research project for machine learning practitioners interested in working with VQ-VAE models. This helps researchers experiment with vector quantized variational autoencoders.

Machine Learning Research Deep Learning Generative Models Variational Autoencoders Neural Networks

Scores updated daily from GitHub, PyPI, and npm data. How scores work