AI Security Research

Original research on LLM jailbreaking, prompt injection, and adversarial AI.