GPT-Jailbreak and gpt_jailbreak_status
These are **complements** — one provides jailbreak techniques/methods while the other tracks their effectiveness and current status across different GPT versions, making them useful together for understanding the landscape of GPT vulnerabilities.
About GPT-Jailbreak
Techiral/GPT-Jailbreak
This repository contains the jailbreaking process for GPT-3, GPT-4, GPT-3.5, ChatGPT, and ChatGPT Plus. By following the instructions in this repository, you will be able to gain access to the inner workings of these language models and modify them to your liking.
This project helps AI safety researchers and red teamers understand and exploit vulnerabilities in large language models like GPT-3, GPT-4, and ChatGPT. It provides specific prompts and instructions that, when entered into these models, can bypass their safety features. The output is an LLM that behaves outside its intended guardrails, which is useful for security analysis or exploring model limitations.
About gpt_jailbreak_status
tg12/gpt_jailbreak_status
This is a repository that aims to provide updates on the status of jailbreaking the OpenAI GPT language model.
This project keeps you informed about the current ability to bypass the safety features of OpenAI's GPT models. It takes in various methods and attempts to 'jailbreak' the AI, and outputs whether those methods are currently working or have been patched. It's for anyone interested in the boundaries and limitations of large language models, particularly AI researchers, ethicists, and prompt engineers exploring AI behavior.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work