Ki6an/fastT5

⚡ boost inference speed of T5 models by 5x & reduce the model size by 3x.

54
/ 100
Established

When performing natural language tasks like summarization, translation, or question answering using T5 models, processing times can be slow, especially with larger models. This tool optimizes your existing T5 models to run up to 5 times faster while reducing their size by 3 times. This is perfect for machine learning engineers or data scientists deploying T5 models in production environments where speed and efficiency are crucial.

589 stars. No commits in the last 6 months. Available on PyPI.

Use this if you need to speed up the inference of your T5-based natural language processing models and reduce their memory footprint.

Not ideal if you are not working with T5 models or if your primary concern is model training speed rather than inference performance.

natural-language-processing machine-learning-operations text-summarization machine-translation question-answering
Stale 6m No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 19 / 25

How are scores calculated?

Stars

589

Forks

74

Language

Python

License

Apache-2.0

Last pushed

Apr 24, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/Ki6an/fastT5"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.