AI-Hypercomputer/maxtext

A simple, performant and scalable Jax LLM!

82
/ 100
Verified

MaxText helps AI engineers and researchers efficiently train and fine-tune large language models (LLMs) on powerful hardware like Google Cloud TPUs and GPUs. You provide raw text data and choose from a library of existing model architectures like Gemma or Llama. MaxText then outputs a highly optimized, custom-trained LLM ready for integration into your applications or further research.

2,169 stars. Actively maintained with 321 commits in the last 30 days. Available on PyPI.

Use this if you need to pre-train or fine-tune large language models from scratch or adapt existing ones for specific tasks, aiming for high performance and scalability on accelerator hardware.

Not ideal if you're looking for an off-the-shelf API for LLM inference or if you don't have access to specialized AI accelerator hardware for training.

large-language-model-training deep-learning-research ai-model-customization high-performance-computing machine-learning-engineering
No Dependents
Maintenance 22 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 25 / 25

How are scores calculated?

Stars

2,169

Forks

485

Language

Python

License

Apache-2.0

Last pushed

Mar 13, 2026

Commits (30d)

321

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AI-Hypercomputer/maxtext"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.