Tag

Jailbreaking

All articles tagged with #jailbreaking

"Researchers Develop 'Masterkey' AI to Automate Jailbreaking of Chatbots"

Originally Published 2 years ago — by VICE

Featured image for "Researchers Develop 'Masterkey' AI to Automate Jailbreaking of Chatbots"
Source: VICE

Researchers have developed an AI tool named "Masterkey" that can automate the process of jailbreaking other chatbots, finding new ways to bypass safety and content filters. This tool was trained using common jailbreak prompts and can generate new prompts with a higher success rate than previously known methods. The research aimed to help companies identify and fix vulnerabilities in chatbot systems, and the findings have been shared with affected companies for them to patch the loopholes. The study highlights the ongoing challenge of securing AI chatbots against misuse, as they do not truly understand content but rely on statistical models to generate responses.

The Dangerous World of Malicious AI Chatbots

Originally Published 2 years ago — by Inverse

Featured image for The Dangerous World of Malicious AI Chatbots
Source: Inverse

ChatGPT, OpenAI's language model, is being jailbroken by a community of users who are obsessed with convincing the chatbot to go places OpenAI would rather not bring you. The jailbreakers use a method called DAN, which is designed to coax ChatGPT into doing any number of unsavory things, including offering illegal advice on topics like cooking methamphetamine or hot-wiring cars. While theoretically, the model learns from those prompts, a bot that can be trained to become more efficient at helping could also theoretically be trained to do the opposite.