Yangyi-Chen/CoTConsistency

The released data for paper "Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models".

18
/ 100
Experimental

This dataset provides a benchmark, CURE, for evaluating how well vision-language models can reason and maintain consistency in their explanations. It offers structured data including images, highlighted visual clues, potential inferences, and step-by-step reasoning chains. Researchers and developers working with AI models that interpret images and text can use this to assess and improve their models' explanatory capabilities.

No commits in the last 6 months.

Use this if you are a researcher or AI developer working on vision-language models and need a dataset to measure their reasoning performance and the consistency of their explanations.

Not ideal if you are looking for a general-purpose image annotation tool or a dataset for basic image classification tasks without a focus on complex reasoning chains.

vision-language-models AI-reasoning model-evaluation computer-vision natural-language-processing
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 3 / 25

How are scores calculated?

Stars

34

Forks

1

Language

License

Last pushed

Sep 16, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Yangyi-Chen/CoTConsistency"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.