L-Zhe/FasySeq
A fast and easy implementation of Transformer with PyTorch.
This toolkit helps machine learning researchers and developers quickly build and train sequence-to-sequence models, especially for tasks like machine translation. You provide pairs of input and output sequences (e.g., sentences in two different languages) and it produces a trained model that can translate new input sentences. It's designed for those who work with natural language processing and need efficient, customizable sequence modeling.
No commits in the last 6 months.
Use this if you are a researcher or developer focused on natural language processing and need a fast, easy-to-modify toolkit for building and experimenting with Transformer-based sequence-to-sequence models.
Not ideal if you are looking for a pre-packaged, out-of-the-box solution without any coding, or if your primary focus is on sequence tasks outside of language, such as time series forecasting or protein folding.
Stars
7
Forks
1
Language
Python
License
Apache-2.0
Category
Last pushed
Aug 31, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/L-Zhe/FasySeq"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
lvapeab/nmt-keras
Neural Machine Translation with Keras
dair-ai/Transformers-Recipe
🧠A study guide to learn about Transformers
SirawitC/Transformer_from_scratch_pytorch
Build a transformer model from scratch using pytorch to understand its inner workings and gain...
jaketae/ensemble-transformers
Ensembling Hugging Face transformers made easy
lof310/transformer
PyTorch implementation of the current SOTA Transformer. Configurable, efficient, and...