CLUEbenchmark/CLUE

中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard

49
/ 100
Emerging

This project offers a comprehensive benchmark for evaluating and comparing the performance of Chinese language understanding models. It provides standardized datasets for various tasks, pre-trained models, and a public leaderboard. This helps researchers and AI developers assess the capabilities of different models on real-world Chinese text understanding challenges.

4,237 stars.

Use this if you are developing or researching AI models for Chinese language processing and need a standardized way to evaluate their performance on tasks like text classification, natural language inference, and reading comprehension.

Not ideal if you are looking for an off-the-shelf application to directly solve a business problem or if your primary focus is on languages other than Chinese.

Chinese-NLP AI-model-evaluation text-understanding natural-language-processing AI-research
No License No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 21 / 25

How are scores calculated?

Stars

4,237

Forks

546

Language

Python

License

Last pushed

Feb 06, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/CLUEbenchmark/CLUE"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.