ChristophReich1996/HyperMixer
PyTorch reimplementation of the paper "HyperMixer: An MLP-based Green AI Alternative to Transformers" [arXiv 2022].
This is a tool for machine learning researchers and practitioners who are building neural networks. It provides an alternative architectural component to the commonly used 'Transformer' blocks, using simpler Multi-Layer Perceptrons (MLPs). You would use this by integrating the HyperMixer block into your PyTorch models, receiving a processed tensor as output that represents a more energy-efficient computation.
No commits in the last 6 months.
Use this if you are a machine learning researcher or engineer exploring novel, more energy-efficient neural network architectures for sequence data processing.
Not ideal if you are looking for a high-level, ready-to-use model for a specific task rather than an architectural component for deep learning.
Stars
18
Forks
—
Language
Python
License
MIT
Category
Last pushed
Mar 28, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/ChristophReich1996/HyperMixer"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Westlake-AI/openmixup
CAIRI Supervised, Semi- and Self-Supervised Visual Representation Learning Toolbox and Benchmark
YU1ut/MixMatch-pytorch
Code for "MixMatch - A Holistic Approach to Semi-Supervised Learning"
kamata1729/QATM_pytorch
Pytorch Implementation of QATM:Quality-Aware Template Matching For Deep Learning
nttcslab/msm-mae
Masked Spectrogram Modeling using Masked Autoencoders for Learning General-purpose Audio Representations
rgeirhos/generalisation-humans-DNNs
Data, code & materials from the paper "Generalisation in humans and deep neural networks" (NeurIPS 2018)