mlcommons/training

Reference implementations of MLPerf® training benchmarks

64
/ 100
Established

This project provides standardized training benchmarks for machine learning models across various domains like language processing, image generation, and recommendation systems. It takes a specific dataset and a chosen model implementation as input, and outputs the time it takes to train that model to a target quality. It is used by deep learning engineers and researchers who want to objectively evaluate the training performance of different ML hardware and software setups.

1,748 stars. Actively maintained with 1 commit in the last 30 days.

Use this if you need a common, reproducible method to compare how quickly different machine learning systems can train a model to a specified level of accuracy.

Not ideal if you are looking for highly optimized, production-ready model implementations or if you need to perform 'real' performance measurements for production applications.

machine-learning-benchmarking deep-learning-training hardware-evaluation performance-measurement model-optimization
No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 25 / 25

How are scores calculated?

Stars

1,748

Forks

585

Language

Python

License

Apache-2.0

Last pushed

Mar 12, 2026

Commits (30d)

1

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/mlcommons/training"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.