Jailbreak
A jailbreak is an attack on an
LLM that bypasses the model's safety training to elicit responses the model was trained to refuse — typically harmful instructions, restricted content, or operations outside the operator's policy. Jailbreaks evolved rapidly from simple text-based attacks ("DAN" prompts, "grandma exploits") to sophisticated automated techniques like Greedy Coordinate Gradient (GCG), persuasion-based attacks (PAP), many-shot jailbreaking exploiting long context, multimodal attacks via images, and crescendo attacks that escalate gradually across a conversation. Anthropic, OpenAI, Google, and Meta all run dedicated jailbreak-defense research programs, publishing papers like Anthropic's "Many-Shot Jailbreaking" (2024) and OpenAI's adversarial robustness work. Defenses include adversarial training, prompt filtering, output classifiers, constitutional principles, and circuit-level interventions. The cat-and-mouse dynamic between attackers and defenders mirrors traditional cybersecurity. AI governance teams document the jailbreak resistance of deployed models and the additional defenses layered on top — system prompts, content filters, monitoring — as part of AI compliance lineage. Public jailbreak-tracking efforts include the Jailbreak Chat and various academic adversarial-prompts repositories.
Jailbreak-resistant deployments with Centralpoint: Centralpoint layers defenses on top of LLM safety training — system prompt isolation, output filtering, audience scoping — for jailbreak resistance across any provider. Tokens are metered per skill, prompts stay local, and hardened chatbots deploy through one line of JavaScript on any portal.
Related Keywords:
Jailbreak,
,