← Back to Glossary
LLM Jailbreak
A technique that circumvents the safety restrictions of large language models to produce outputs the model was designed to refuse.
A technique that circumvents the safety restrictions of large language models to produce outputs the model was designed to refuse.