zihao-ai/unthinking_vulnerability
To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning Models
This project identifies and explores a critical 'Unthinking Vulnerability' in Large Reasoning Models (LRMs), where specific inputs can bypass their intended reasoning processes. It provides tools to either maliciously exploit this flaw to induce incorrect outputs or beneficially monitor and enhance the safety and efficiency of these models. AI developers and researchers working with LRMs would use this to build more robust and secure AI systems.
No commits in the last 6 months.
Use this if you are an AI developer or researcher concerned about the reliability and security of Large Reasoning Models and need tools to test for and mitigate reasoning bypass vulnerabilities.
Not ideal if you are looking for an end-user application or a pre-built solution for a specific business problem, as this is a research toolkit for model developers.
Stars
33
Forks
—
Language
Python
License
—
Category
Last pushed
May 21, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/agents/zihao-ai/unthinking_vulnerability"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Ed1s0nZ/CyberStrikeAI
CyberStrikeAI is an AI-native security testing platform built in Go. It integrates 100+ security...
GH05TCREW/pentestagent
PentestAgent is an AI agent framework for black-box security testing, supporting bug bounty,...
vxcontrol/pentagi
✨ Fully autonomous AI Agents system capable of performing complex penetration testing tasks
asaotomo/FofaMap
FofaMap v2.0 是一款基于 Python3 开发的全网首个 AI 驱动红队资产测绘智能体。在延续原有 FOFA 数据采集、存活检测、统计聚合、图标 Hash...
SanMuzZzZz/LuaN1aoAgent
LuaN1aoAgent is a cognitive-driven AI hacker. It is a fully autonomous AI penetration testing...