Glossary
Jailbreak (AI Guardrail Bypass)
A technique aimed at circumventing the restrictions and guardrails of an LLM to make it produce forbidden or dangerous content. Jailbreaks exploit flaws in the model's alignment through carefully crafted prompts. For businesses, a successful jailbreak on a chatbot can expose internal data or produce harmful responses.