tommyip/mamba2-minimal
Minimal Mamba-2 implementation in PyTorch
This project offers a highly efficient way to build language models for tasks like text generation or sequence processing, without the computational overhead of traditional Transformer models. It takes in sequential data (like text or time series) and processes it into output logits, enabling rapid training and constant-time inference, especially useful for very long sequences. It's designed for machine learning practitioners and researchers working with sequential data.
243 stars. No commits in the last 6 months.
Use this if you need to develop or experiment with cutting-edge foundation models for sequential data that are faster and more memory-efficient than Transformer architectures, particularly for long sequences.
Not ideal if you are looking for a pre-trained, ready-to-use application and not a foundational building block for model development.
Stars
243
Forks
16
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 17, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tommyip/mamba2-minimal"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ZHZisZZ/dllm
dLLM: Simple Diffusion Language Modeling
pengzhangzhi/Open-dLLM
Open diffusion language model for code generation — releasing pretraining, evaluation,...
EnnengYang/Awesome-Model-Merging-Methods-Theories-Applications
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. ACM...
THUDM/LongWriter
[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
AIoT-MLSys-Lab/SVD-LLM
[ICLR 2025🔥] SVD-LLM & [NAACL 2025🔥] SVD-LLM V2