A New Trick Uses AI to Jailbreak AI Models—Including GPT-4

11/12/2023 5 min
A New Trick Uses AI to Jailbreak AI Models—Including GPT-4

Listen "A New Trick Uses AI to Jailbreak AI Models—Including GPT-4"

Episode Synopsis


Adversarial algorithms can systematically probe large language models like OpenAI’s GPT-4 for weaknesses that can make them misbehave. Read the story here.

Learn about your ad choices: dovetail.prx.org/ad-choices