google-deepmind/optax
Optax is a gradient processing and optimization library for JAX.
This library helps machine learning researchers efficiently train their neural networks. It takes a JAX-based model's parameters and gradients, applies various optimization techniques, and outputs updated parameters to improve model performance. It's used by machine learning practitioners and researchers who build and experiment with custom deep learning models.
2,207 stars. Used by 41 other packages. Actively maintained with 11 commits in the last 30 days. Available on PyPI.
Use this if you are a machine learning researcher building neural networks in JAX and need a flexible toolkit to customize your gradient processing and optimization algorithms.
Not ideal if you are looking for a high-level, opinionated deep learning framework that handles model architecture and training loops for you.
Stars
2,207
Forks
318
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 12, 2026
Commits (30d)
11
Dependencies
4
Reverse dependents
41
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/google-deepmind/optax"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Related frameworks
explosion/thinc
🔮 A refreshing functional take on deep learning, compatible with your favorite libraries
patrick-kidger/diffrax
Numerical differential equation solvers in JAX. Autodifferentiable and GPU-capable....
google/grain
Library for reading and processing ML training data.
patrick-kidger/equinox
Elegant easy-to-use neural networks + scientific computing in JAX. https://docs.kidger.site/equinox/
extropic-ai/thrml
Thermodynamic Hypergraphical Model Library in JAX