myscience/x-lstm
Pytorch implementation of the xLSTM model by Beck et al. (2024)
This project provides an unofficial implementation of the xLSTM model, a modern alternative to Transformers and State-Space models for tasks involving sequential data like text. It allows researchers and practitioners to train and experiment with xLSTM for generating text, taking raw text as input and producing new, coherent text sequences. It's designed for machine learning researchers and AI practitioners exploring advanced recurrent neural network architectures.
183 stars. No commits in the last 6 months.
Use this if you are a machine learning researcher or practitioner interested in experimenting with cutting-edge Long Short-Term Memory architectures for natural language processing tasks, particularly text generation, and want a Pytorch-based implementation with multi-GPU training support.
Not ideal if you need a production-ready, highly optimized, and officially supported implementation of xLSTM, or if your primary focus is on immediate application rather than model exploration and understanding.
Stars
183
Forks
19
Language
Python
License
—
Category
Last pushed
Aug 12, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/myscience/x-lstm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
mosaicml/llm-foundry
LLM training code for Databricks foundation models
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소