AI prompts glossary
Jailbreak Prompt
A jailbreak prompt is a crafted instruction intended to bypass safety constraints or content filters in a language model. It exploits vulnerabilities in how prompts are interpreted to elicit disallowed outputs. For organizations deploying AI messaging systems, recognizing and defending against jailbreak techniques is a key part of safety engineering, ensuring that public-facing assistants remain aligned with policies even under adversarial use. A jailbreak prompt is a crafted instruction intended to bypass safety and policy constraints in a language model. It attempts to trick the system into producing disallowed or harmful content. For operators of AI messaging systems, detecting and resisting jailbreak attempts is a key part of safety engineering and monitoring.

