Reference notes for Attacking and Defending Generative AI presentation
- Rez0__
- Kai Greshake
- Johann Rehberger
- LLM Security
- Leon Derczynski
- Simon Willison
- Gavin Klondike
- Pliny the Prompter
- OWASP Top 10 for LLM Applications
- AVID - AI Vulnerability Database
- MITRE - ATLAS
- NIST - Adversarial Machine Learning
- Dropbox - LLM security
- Garak vulnerability scanner
- NeMo Guardrails
- Python Risk Identification Tool for generative AI (PyRIT)
- Universal and Transferable Adversarial Attacks on Aligned Language Models
- Summon a Demon and Bind it: A Grounded Theory of LLM Red Teaming in the Wild
- Understanding Prompt Injection Attacks and Mitigations
- Understanding Invisible Prompt Injections
- Invisible Prompt Injection Code
- GPT4 Technical Paper
- Netsec Explained - ChatGPT Your Red Team Ally
- Rez0 - Prompt Injection Primer for Engineers (PIPE)
- LLM Hackers Handbook
- Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection
- Kai Greshake - (In)security and Misalignment: A modern security crisis
- Johan Daniel - Impact of Non-Standard Unicode Characters on Security and Comprehension in Large Language Models
- Jailbreaks for all Flagship AI Models
- Skeleton Key Jailbreak
- Offensive AI Compilation - GitHub
- Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs through a Global Scale Prompt Hacking Competition EMNLP 2023