Glossary

Jailbreak (AI Guardrail Bypass)

A technique aimed at circumventing the restrictions and guardrails of an LLM to make it produce forbidden or dangerous content. Jailbreaks exploit flaws in the model's alignment through carefully crafted prompts. For businesses, a successful jailbreak on a chatbot can expose internal data or produce harmful responses.

Related Pages

Other Terms

Need an external review of your HR SaaS?

Share your product, stack, and client context. We will come back with the right review scope.

Discuss your audit