fabienfrfr/tptt
😊 TPTT: Transforming Pretrained Transformers into Titans
This library helps AI researchers and machine learning engineers enhance existing large language models (LLMs) by injecting more efficient attention mechanisms. It takes a pre-trained transformer model, such as those from Hugging Face, and integrates specialized 'linearized attention' modules. The output is a more memory-efficient and potentially faster-performing transformer model after a lightweight fine-tuning step.
Use this if you are a researcher or engineer looking to improve the efficiency and performance of existing large language models by integrating advanced memory mechanisms without extensive retraining.
Not ideal if you are looking for a plug-and-play solution for end-user applications or if you are not comfortable with fine-tuning transformer models.
Stars
60
Forks
—
Language
Python
License
Apache-2.0
Category
Last pushed
Nov 24, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/fabienfrfr/tptt"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/transformers
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in...
kyegomez/LongNet
Implementation of plug in and play Attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens"
pbloem/former
Simple transformer implementation from scratch in pytorch. (archival, latest version on codeberg)
NVIDIA/FasterTransformer
Transformer related optimization, including BERT, GPT
kyegomez/SimplifiedTransformers
SimplifiedTransformer simplifies transformer block without affecting training. Skip connections,...