ztjhz/t5-jax

JAX implementation of the T5 model: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

22
/ 100
Experimental

This is an enhanced implementation of the T5 model, a powerful AI model for text-to-text tasks like summarization, translation, or question answering. It processes input text to generate relevant output text, making it useful for AI researchers and engineers who are building and experimenting with large language models. The project aims to provide a clearer and more performant T5 codebase, especially when working with Google Cloud TPUs.

No commits in the last 6 months.

Use this if you are an AI researcher or engineer working with Transformer-based language models and need a high-performance, clear, and educational T5 implementation, particularly for Google Cloud TPUs.

Not ideal if you are a practitioner looking for a ready-to-use application of T5 without diving into its underlying architecture and technical optimizations.

natural-language-processing large-language-models text-generation machine-translation AI-model-development
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 0 / 25

How are scores calculated?

Stars

24

Forks

Language

Python

License

MIT

Last pushed

Jun 10, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/ztjhz/t5-jax"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.