Awesome-Jailbreak-on-LLMs and awesome-llm-jailbreaks
These are complements that serve different purposes: one is an academic/research repository cataloging jailbreak methodologies with papers and analyses, while the other is a practical payload collection for testing and exploitation, used together to understand jailbreaks both theoretically and operationally.
About Awesome-Jailbreak-on-LLMs
yueliu1999/Awesome-Jailbreak-on-LLMs
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses.
This resource provides a curated collection of techniques for evaluating and improving the safety of Large Language Models (LLMs). It includes research papers, code, and datasets related to both 'jailbreak' attacks (attempts to bypass safety mechanisms) and defenses against them. AI safety researchers and practitioners who are building or deploying LLMs would use this to understand vulnerabilities and develop more robust, responsible AI systems.
About awesome-llm-jailbreaks
Techiral/awesome-llm-jailbreaks
Latest AI Jailbreak Payloads & Exploit Techniques for GPT, QWEN, and all LLM Models
Scores updated daily from GitHub, PyPI, and npm data. How scores work