kyegomez/MambaTransformer
Integrating Mamba/SSMs with Transformer for Enhanced Long Context and High-Quality Sequence Modeling
This project helps AI developers build advanced language models that can understand and generate very long sequences of text more effectively. It takes in raw text or tokenized sequences and outputs predictions or generated text, suitable for tasks requiring deep understanding of extensive content. Developers working on sophisticated natural language processing applications will find this useful.
215 stars. Available on PyPI.
Use this if you are developing AI models that need to process or generate very long texts with high accuracy and improved reasoning, such as in advanced content generation or complex document analysis.
Not ideal if you are working with short, simple text sequences or if you prefer a standard, widely adopted transformer architecture for typical NLP tasks.
Stars
215
Forks
16
Language
Python
License
MIT
Category
Last pushed
Jan 30, 2026
Commits (30d)
0
Dependencies
3
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/kyegomez/MambaTransformer"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related models
kyegomez/MambaByte
Implementation of MambaByte in "MambaByte: Token-free Selective State Space Model" in Pytorch and Zeta
kyegomez/HSSS
Implementation of a Hierarchical Mamba as described in the paper: "Hierarchical State Space...
ghostperpper007/small_programming_model
A from-scratch Python code model with GNN-based structure encoding, Mamba-style SSM decoding,...