Jailbreak
-
Blog
Time Bandit ChatGPT jailbreak bypasses safeguards on sensitive topics
A ChatGPT jailbreak flaw, dubbed “Time Bandit,” allows you to bypass OpenAI’s safety guidelines when asking for detailed instructions on sensitive topics, including the creation of weapons, information on nuclear topics, and malware creation. The vulnerability was discovered by cybersecurity and AI researcher David Kuszmar, who found that ChatGPT suffered from “temporal confusion,” making it possible to put the LLM…
Read More » -
Blog
20% of Generative AI ‘Jailbreak’ Attacks are Successful
Generative AI jailbreak attacks, where models are instructed to ignore their safeguards, succeed 20% of the time, research has found. On average, adversaries need just 42 seconds and five interactions to break through. In some cases, attacks occur in as little as four seconds. These findings both highlight the significant vulnerabilities in current GenAI algorithms and the difficulty in preventing…
Read More »