soskek/attention_is_all_you_need
Transformer of "Attention Is All You Need" (Vaswani et al. 2017) by Chainer.
This project helps machine learning engineers or researchers implement the Transformer model for sequence-to-sequence tasks. You input pairs of text in different languages, and it generates a trained model capable of translating or transforming new text sequences. It's designed for those who need to experiment with or apply the Transformer architecture to problems like machine translation or text summarization.
323 stars. No commits in the last 6 months.
Use this if you need a Python-based implementation of the Transformer model using Chainer for sequence-to-sequence problems, particularly if you are experimenting with different training strategies or model configurations.
Not ideal if you need a production-ready, highly optimized translation system or if you are not comfortable working with machine learning model implementations and datasets directly.
Stars
323
Forks
66
Language
Jupyter Notebook
License
BSD-3-Clause
Category
Last pushed
Oct 03, 2017
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/soskek/attention_is_all_you_need"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
philipperemy/keras-attention
Keras Attention Layer (Luong and Bahdanau scores).
tatp22/linformer-pytorch
My take on a practical implementation of Linformer for Pytorch.
ematvey/hierarchical-attention-networks
Document classification with Hierarchical Attention Networks in TensorFlow. WARNING: project is...
datalogue/keras-attention
Visualizing RNNs using the attention mechanism
thushv89/attention_keras
Keras Layer implementation of Attention for Sequential models