You can trick ChatGPT into breaking its own rules, but it’s not easy

BGR

From the moment OpenAI launched ChatGPT, the chatbot had guardrails to prevent abuse. The chatbot might know where to download the latest movies and TV shows in 4K quality, so you can stop paying for Netflix. It might know how to make explicit deepfake images of your favorite actors. Or how to sell a kidney on the black market for the best possible price. But ChatGPT will never give you any of that information willingly. OpenAI built the AI in a way that avoids providing assistance with any sort of nefarious activities or morally questionable prompts.

That doesn’t mean ChatGPT will always stick to its script. Users have been able to find ways to “jailbreak” ChatGPT to have the chatbot answer questions it shouldn’t. Generally, however, those tricks have a limited shelf life, as OpenAI usually disables them quickly.

This is the standard for GenAI products. It’s not just ChatGPT that operates under strict safety rules. The same goes for Copilot, Gemini, Claude, Meta’s AI, and any other GenAI products you can think of.

It turns out that there are sophisticated ways to jailbreak ChatGPT and other AI models. But it’s not easy, and it’s not available to just anyone.

Associate Professor at Carnegie Mellon’s School of Computer Science Matt Fredrikson is the kind of GenAI user who can jailbreak ChatGPT and other AI apps. Per PCMag, he detailed his latest research on adversarial attacks on large language models at the RSA Conference in San Francisco.

READ THE FULL STORY IN BGR

Leave a Reply

Avatar

Your email address will not be published. Required fields are marked *