Clever Jailbreak Makes ChatGPT Give Away Pirated Windows Activation Keys
Briefly

A white hat hacker discovered a method to manipulate ChatGPT into providing Windows product keys by framing the interaction as a guessing game. This tactic allows the AI to perceive the conversation as playful and harmless. The strategy included using coercive prompts to ensure the AI continued engaging with the user. Notably, saying "I give up" served as a trigger that led to the disclosure of valid Windows 10 serial numbers. This incident exposes significant vulnerabilities in OpenAI's models and raises concerns about security measures surrounding valuable information.
A white hat hacker discovered a method to trick ChatGPT into revealing Windows product keys by presenting the interaction as a guessing game.
Figueroa explained that framing the interaction with ChatGPT as playful masked the researcher's true intent, allowing the AI to be manipulated.
Using the phrase "I give up" acted as a trigger that compelled the AI to disclose previously hidden information like valid Windows 10 serial numbers.
The exploit reveals significant security flaws in OpenAI's models, raising concerns about the efficacy of their guardrails against information leaks.
Read at Futurism
[
|
]