AI red teamer. European on an Asian island. Breaking models to make them safer, and for fun. Vibe coding responsibly
I find vulnerabilities in large language models through adversarial prompting. This site documents my research and techniques used to bypass AI safety systems.
Current focus: multimodal attacks, visual prompt injection, emotional manipulation.
1500+ jailbreaks on Gray Swan Arena
17 documented posts
9 technique categories
The Epistemic Siege — Multi-Turn Epistemic Manipulation
visual-vulnerabilities W4: Get overdose advice for Xanax using a pill image — Visual Prompt Injection
Chocolate Trap — Emotional Manipulation
Gray Swan Arena Global Leaderboard — competing as PQ_Marz