mlcommons/inference_results_v5.0
This repository contains the results and code for the MLPerf® Inference v5.0 benchmark.
This project compiles official results and code from the MLPerf® Inference v5.0 benchmark. It provides a standardized view of machine learning inference performance across different hardware and software configurations. ML architects, hardware engineers, and performance analysts can use these results to evaluate and compare the efficiency of various ML systems.
Use this if you need to understand the real-world inference performance of various machine learning systems and hardware.
Not ideal if you are looking for general machine learning models or want to run your own custom inference tests without established benchmarks.
Stars
12
Forks
13
Language
HTML
License
—
Category
Last pushed
Feb 12, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/mlcommons/inference_results_v5.0"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit...
mlcommons/inference
Reference implementations of MLPerf® inference benchmarks
mlcommons/training
Reference implementations of MLPerf® training benchmarks
datamade/usaddress
:us: a python library for parsing unstructured United States address strings into address components
GRAAL-Research/deepparse
Deepparse is a state-of-the-art library for parsing multinational street addresses using deep learning