elbayadm/attn2d
Pervasive Attention: 2D Convolutional Networks for Sequence-to-Sequence Prediction
This project helps machine translation researchers and engineers explore and implement advanced neural machine translation (NMT) models. It takes source language text and produces target language text, focusing on more efficient and sophisticated methods for handling sequences. The primary users are professionals developing or studying NMT systems.
498 stars. No commits in the last 6 months.
Use this if you are a machine translation researcher or engineer looking to experiment with 2D convolutional networks for sequence-to-sequence prediction or efficient wait-k models for simultaneous translation.
Not ideal if you need an out-of-the-box translation service or a simple API for general-purpose text translation without deep model customization.
Stars
498
Forks
71
Language
Python
License
MIT
Category
Last pushed
May 08, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/elbayadm/attn2d"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
xv44586/toolkit4nlp
transformers implement (architecture, task example, serving and more)
luozhouyang/transformers-keras
Transformer-based models implemented in tensorflow 2.x(using keras).
ufal/neuralmonkey
An open-source tool for sequence learning in NLP built on TensorFlow.
graykode/xlnet-Pytorch
Simple XLNet implementation with Pytorch Wrapper
uzaymacar/attention-mechanisms
Implementations for a family of attention mechanisms, suitable for all kinds of natural language...