shammur/SemEval2022Task3

The PreTENS shared task hosted at SemEval 2022 aims at focusing on semantic competence with specific attention on the evaluation of language models with respect to the recognition of appropriate taxonomic relations between two nominal arguments (i.e. cases where one is a supercategory of the other, or in extensional terms, one denotes a superset of the other).

27
/ 100
Experimental

This project helps evaluate how well language models understand subtle semantic relationships in sentences, specifically focusing on whether one noun implies another as a supercategory (e.g., 'dog' implies 'animal'). It takes sentences in English, Italian, or French as input and outputs a judgment on their acceptability or a score indicating the degree of acceptance. Linguists, computational semanticists, or AI researchers focused on natural language understanding would use this.

No commits in the last 6 months.

Use this if you are developing or evaluating language models and need to assess their ability to recognize taxonomic relations and semantic acceptability in various sentence structures.

Not ideal if you need a tool for general text classification or sentiment analysis, as its focus is specifically on nuanced semantic relationships.

computational-linguistics natural-language-understanding semantic-analysis language-model-evaluation AI-research
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

12

Forks

1

Language

Jupyter Notebook

License

MIT

Last pushed

Feb 05, 2022

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/shammur/SemEval2022Task3"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.