cpldcpu/MisguidedAttention

A collection of prompts to challenge the reasoning abilities of large language models in presence of misguiding information

41
/ 100
Emerging

This project offers a collection of specially crafted questions designed to test how well large language models (LLMs) reason when faced with misleading information. It helps evaluate if an LLM can logically solve a problem as stated, or if it defaults to familiar, but incorrect, answers learned during training. The input is a 'trick question' prompt, and the output is the LLM's response, revealing its reasoning strengths and weaknesses. Anyone responsible for evaluating or implementing LLMs in critical applications would find this useful.

466 stars. No commits in the last 6 months.

Use this if you need to rigorously test and benchmark the reasoning and problem-solving capabilities of different large language models, especially when 'common sense' or 'pattern recognition' might lead them astray.

Not ideal if you are looking for a tool to generate prompts for general creative writing tasks or for basic information retrieval from LLMs.

LLM-evaluation AI-benchmarking cognitive-testing prompt-engineering AI-safety
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

466

Forks

27

Language

Python

License

CC0-1.0

Last pushed

Jul 31, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/cpldcpu/MisguidedAttention"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.