OpenNLPLab/Tnn
[ICLR 2023] Official implementation of Transnormer in our ICLR 2023 paper - Toeplitz Neural Network for Sequence Modeling
This project helps researchers and machine learning engineers reproduce results from the ICLR 2023 "Toeplitz Neural Network for Sequence Modeling" paper. It provides scripts and instructions to set up environments, preprocess data like WikiText-103 or LRA datasets, and train language or image models. The primary users are academic researchers or ML engineers focused on replicating or extending cutting-edge sequence modeling techniques.
No commits in the last 6 months.
Use this if you are a machine learning researcher or engineer looking to reproduce or build upon the results of the Toeplitz Neural Network (TNN) paper for sequence modeling tasks like language modeling or image generation.
Not ideal if you are an end-user looking for a pre-trained model or an easy-to-use API for direct application in a business context, as this repository focuses on research reproduction rather than out-of-the-box solutions.
Stars
81
Forks
4
Language
Python
License
—
Category
Last pushed
Apr 24, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/OpenNLPLab/Tnn"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/transformers-bloom-inference
Fast Inference Solutions for BLOOM
Tencent/TurboTransformers
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc)...
mit-han-lab/lite-transformer
[ICLR 2020] Lite Transformer with Long-Short Range Attention
mit-han-lab/hardware-aware-transformers
[ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing
LibreTranslate/Locomotive
Toolkit for training/converting LibreTranslate compatible language models 🚂