A New Trick Uses AI to Jailbreak AI Models—Including GPT-4
Credit to Author: Will Knight| Date: Tue, 05 Dec 2023 11:00:00 +0000
Adversarial algorithms can systematically probe large language models like OpenAI’s GPT-4 for weaknesses that can make them misbehave.
Read more