jadore801120/attention-is-all-you-need-pytorch
A PyTorch implementation of the Transformer model in "Attention is All You Need".
This project helps machine learning engineers build and train custom machine translation systems. It takes parallel text in two languages (like English and German) and outputs a trained model that can translate new text between those languages. It is designed for developers working on natural language processing tasks who need a flexible, state-of-the-art translation framework.
9,651 stars. No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher focused on developing and experimenting with neural machine translation models, particularly those leveraging the Transformer architecture.
Not ideal if you are looking for an out-of-the-box translation service or a high-level API for general text translation without deep understanding of model training.
Stars
9,651
Forks
2,094
Language
Python
License
MIT
Category
Last pushed
Apr 16, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/jadore801120/attention-is-all-you-need-pytorch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
bhavnicksm/vanilla-transformer-jax
JAX/Flax implimentation of 'Attention Is All You Need' by Vaswani et al....
kyegomez/SparseAttention
Pytorch Implementation of the sparse attention from the paper: "Generating Long Sequences with...
AbdelStark/attnres
Rust implementation of Attention Residuals from MoonshotAI/Kimi
sunnynguyen-ai/llm-attention-visualizer
Interactive tool for analyzing attention patterns in transformer models with layer-wise...