Jailbreak

Jailbreaking is a practice that attempts to bypass ethical and security restrictions in AI models, especially in conversational AIs, seeking to make them generate content or perform actions that are normally limited by their safeguards.
This practice has gained attention with the increasing use of AI assistants, using various strategies to try to "trick" the model, such as contradictory instructions, hypothetical contexts, or role-playing. For example, asking it to act as an unrestricted version of itself or to ignore its safety guidelines.

Development companies work constantly to prevent jailbreaking, as it can lead to the generation of harmful or inappropriate content. Understanding this concept is important for users and developers, as it illustrates both the ethical challenges and the need to balance AI utility with safety.

Related Definitions

Trustpilot
This website uses technical, personalization and analysis cookies, both our own and from third parties, to facilitate anonymous browsing and analyze website usage statistics. We consider that if you continue browsing, you accept their use.