Loading...

Jailbreak

An attempt to bypass a model's safety restrictions through crafted prompts. Jailbreak resistance is part of safety evaluation, and successful jailbreaks may trigger AUP violations, incident response obligations, and potential liability for resulting harms.

See: Adversarial attack; Prompt injection; Safety evaluation