dependentsign/Awesome-LLM-based-Evaluators
✨✨Latest Papers about LLM-based Evaluators
This is a curated collection of research papers focused on evaluating large language models (LLMs) using other LLMs, rather than human review. It helps researchers and practitioners understand the latest advancements in automatically assessing LLM performance. The output is a list of relevant academic papers, often with links to their code, for those working on or interested in the quality of AI language models.
Use this if you are a researcher or AI practitioner needing to stay current with academic work on automated LLM evaluation methods.
Not ideal if you are looking for ready-to-use software or a guide on how to implement LLM evaluations yourself, as this primarily lists academic papers.
Stars
32
Forks
1
Language
—
License
—
Category
Last pushed
Feb 26, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dependentsign/Awesome-LLM-based-Evaluators"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
FairyFali/SLMs-Survey
Survey of Small Language Models from Penn State, ...
USC-FORTIS/AD-LLM
[ACL Findings 2025] A benchmark for anomaly detection using large language models. It supports...
swordlidev/Efficient-Multimodal-LLMs-Survey
Efficient Multimodal Large Language Models: A Survey
zabir-nabil/awesome-multilingual-large-language-models
A comprehensive collection of multilingual datasets and large language models, meticulously...
AIoT-MLSys-Lab/Efficient-LLMs-Survey
[TMLR 2024] Efficient Large Language Models: A Survey