SAILResearch/awesome-foundation-model-leaderboards
A curated list of awesome leaderboard-oriented resources for AI domain
This is a curated collection of resources for anyone developing, evaluating, or comparing AI models. It brings together various AI leaderboards, development tools, and evaluation organizations. Researchers and practitioners in AI can use this to find relevant benchmarks and compare the performance of different foundation models across many categories.
321 stars.
Use this if you need to find an existing AI leaderboard, understand how to set up your own model evaluation, or discover tools for managing AI competitions.
Not ideal if you are a general user looking for an introduction to AI concepts or a developer seeking low-level programming libraries for building AI models from scratch.
Stars
321
Forks
36
Language
—
License
—
Category
Last pushed
Mar 09, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/agents/SAILResearch/awesome-foundation-model-leaderboards"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
StonyBrookNLP/appworld
🌍 AppWorld: A Controllable World of Apps and People for Benchmarking Function Calling and...
qualifire-dev/rogue
AI Agent Evaluator & Red Team Platform
microsoft/WindowsAgentArena
Windows Agent Arena (WAA) 🪟 is a scalable OS platform for testing and benchmarking of...
future-agi/ai-evaluation
Evaluation Framework for all your AI related Workflows
agentscope-ai/OpenJudge
OpenJudge: A Unified Framework for Holistic Evaluation and Quality Rewards