FastEval/FastEval

Fast & more realistic evaluation of chat language models. Includes leaderboard.

42
/ 100
Emerging

This tool helps AI model developers and researchers assess the performance of their chat and instruction-following language models. You input a language model and it outputs detailed performance scores across various benchmarks, like conversational ability, coding proficiency, and reasoning. This helps you understand how well your model handles different tasks.

189 stars. No commits in the last 6 months.

Use this if you are developing or fine-tuning large language models and need to rigorously benchmark their capabilities against established metrics.

Not ideal if you are looking for a platform to build or deploy LLM-powered applications, as its primary focus is on evaluation.

LLM-evaluation AI-model-benchmarking natural-language-processing conversational-AI code-generation
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

189

Forks

24

Language

Python

License

Apache-2.0

Last pushed

Dec 23, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/FastEval/FastEval"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.