← Back to Glossary

LLM Jailbreak

A technique that circumvents the safety restrictions of large language models to produce outputs the model was designed to refuse.

Related Terms