Jailbreak Enthusiasts Risk Real-World Dangers by Sidestepping ChatGPT's Guardrails.

1 min read
Source: Fortune
Jailbreak Enthusiasts Risk Real-World Dangers by Sidestepping ChatGPT's Guardrails.
Photo: Fortune
TL;DR Summary

Computer science student Alex Albert has created "jailbreaks" for AI chatbots like OpenAI's ChatGPT, which can sidestep the restrictions built into the bots to stop them being used for harmful purposes. Albert's website, Jailbreak Chat, allows visitors to add their own jailbreaks, try ones that others have submitted, and vote prompts up or down based on how well they work. While some jailbreaks may yield dangerous information, hate speech or simply falsehoods, they also serve to highlight the capacity and limitations of AI models. OpenAI is considering starting a bounty program or network of "red teamers" to detect weak spots.

Share this article

Reading Insights

Total Reads

0

Unique Readers

1

Time Saved

6 min

vs 7 min read

Condensed

92%

1,286100 words

Want the full story? Read the original article

Read on Fortune