cloudmercato/ollama-benchmark

Handy tool to measure the performance and efficiency of LLMs workloads.

39
/ 100
Emerging

This tool helps AI engineers and researchers assess how well their Ollama-hosted large language models (LLMs) are performing. It takes various LLM models and test parameters as input and outputs detailed performance metrics like response speed, embedding generation time, and even the quality of answers. You can use it to compare different models or optimize a single model's setup for specific tasks.

No commits in the last 6 months.

Use this if you need to systematically measure and compare the speed, efficiency, and quality of different LLM configurations running on Ollama.

Not ideal if you are looking for a tool to deploy or manage your LLMs, or if you need to benchmark models hosted on platforms other than Ollama.

LLM-benchmarking model-evaluation AI-performance natural-language-processing generative-AI
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 12 / 25

How are scores calculated?

Stars

76

Forks

8

Language

Python

License

MIT

Last pushed

Apr 26, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/cloudmercato/ollama-benchmark"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.