tommyip/mamba2-minimal

Minimal Mamba-2 implementation in PyTorch

37
/ 100
Emerging

This project offers a highly efficient way to build language models for tasks like text generation or sequence processing, without the computational overhead of traditional Transformer models. It takes in sequential data (like text or time series) and processes it into output logits, enabling rapid training and constant-time inference, especially useful for very long sequences. It's designed for machine learning practitioners and researchers working with sequential data.

243 stars. No commits in the last 6 months.

Use this if you need to develop or experiment with cutting-edge foundation models for sequential data that are faster and more memory-efficient than Transformer architectures, particularly for long sequences.

Not ideal if you are looking for a pre-trained, ready-to-use application and not a foundational building block for model development.

natural-language-processing machine-learning-research sequence-modeling deep-learning-optimization
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

243

Forks

16

Language

Python

License

Apache-2.0

Last pushed

Jun 17, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tommyip/mamba2-minimal"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.