SapienzaNLP/ita-bench

A collection of Italian benchmarks for LLM evaluation

32
/ 100
Emerging

ITA-Bench helps researchers and developers evaluate how well Large Language Models (LLMs) understand and generate Italian. You input an LLM and it produces performance scores across various tasks like question answering, commonsense reasoning, and named entity recognition in Italian. This tool is for AI researchers, natural language processing engineers, and data scientists working with Italian language models.

Use this if you need a standardized way to measure the capabilities of Italian LLMs across diverse linguistic tasks and benchmarks.

Not ideal if your primary focus is on English LLM evaluation or if you need to create entirely new Italian evaluation datasets from scratch.

Italian NLP LLM evaluation natural language processing AI model benchmarking computational linguistics
No Package No Dependents
Maintenance 6 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 3 / 25

How are scores calculated?

Stars

37

Forks

1

Language

Python

License

Apache-2.0

Last pushed

Dec 02, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/SapienzaNLP/ita-bench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.