IAAR-Shanghai/NewsBench

[ACL 2024 Main] NewsBench: A Systematic Evaluation Framework for Assessing Editorial Capabilities of Large Language Models in Chinese Journalism

26
/ 100
Experimental

This project helps Chinese journalism professionals evaluate how well large language models (LLMs) perform editorial tasks like summarization or headline generation, and if they adhere to safety guidelines. It takes a Chinese news article or topic as input and provides an assessment of the LLM's journalistic writing proficiency and safety adherence. News editors, content strategists, and researchers in Chinese media would use this to ensure AI-generated content meets industry standards.

No commits in the last 6 months.

Use this if you need to systematically test and compare the editorial capabilities and safety of different large language models for Chinese news content creation.

Not ideal if your focus is on evaluating LLMs for languages other than Chinese or for tasks outside of journalistic editorial workflows.

Chinese-journalism news-editing AI-content-evaluation media-ethics large-language-models
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 3 / 25

How are scores calculated?

Stars

34

Forks

1

Language

Python

License

Apache-2.0

Last pushed

Jun 25, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/IAAR-Shanghai/NewsBench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.