dependentsign/Awesome-LLM-based-Evaluators

✨✨Latest Papers about LLM-based Evaluators

29
/ 100
Experimental

This is a curated collection of research papers focused on evaluating large language models (LLMs) using other LLMs, rather than human review. It helps researchers and practitioners understand the latest advancements in automatically assessing LLM performance. The output is a list of relevant academic papers, often with links to their code, for those working on or interested in the quality of AI language models.

Use this if you are a researcher or AI practitioner needing to stay current with academic work on automated LLM evaluation methods.

Not ideal if you are looking for ready-to-use software or a guide on how to implement LLM evaluations yourself, as this primarily lists academic papers.

AI-evaluation natural-language-processing-research large-language-models AI-benchmarking machine-learning-research
No License No Package No Dependents
Maintenance 10 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 4 / 25

How are scores calculated?

Stars

32

Forks

1

Language

License

Last pushed

Feb 26, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dependentsign/Awesome-LLM-based-Evaluators"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.