Jailbreaking AI - Graph View Techniques used to bypass an AI model's safety guardrails and restrictions to produce outputs it was designed to refuse. View concept details Related ConceptsPrompt Injection AI Guardrails AI Safety AI Alignment Constitutional AI Reinforcement Learning from Human Feedback (RLHF) Red Teaming Prompt Adherence ← Back to full graph