CLUEbenchmark/SuperCLUE
SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese
Need to understand how well large language models (LLMs) perform on tasks relevant to the Chinese language and culture? SuperCLUE provides comprehensive evaluations across key capabilities like language understanding, professional knowledge, and AI agent performance. It helps compare different Chinese LLMs and understand their strengths and weaknesses, offering a ranked list of models and detailed breakdowns of their abilities.
3,277 stars.
Use this if you need to compare or select Chinese large language models based on their performance across a wide range of practical applications and specialized skills.
Not ideal if you are looking for an evaluation of non-Chinese language models or highly specialized, niche technical benchmarks not related to general LLM capabilities.
Stars
3,277
Forks
112
Language
—
License
—
Category
Last pushed
Feb 06, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/CLUEbenchmark/SuperCLUE"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AI-Planning/l2p
Library for LLM-driven action model acquisition via natural language
datawhalechina/self-llm
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
microsoft/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
theaniketgiri/create-llm
The fastest way to build and start training your own LLM. CLI tool that scaffolds...