Article 6GX9G A New Trick Uses AI to Jailbreak AI Models—Including GPT-4

A New Trick Uses AI to Jailbreak AI Models—Including GPT-4

by
Will Knight
from Feed: All Latest on (#6GX9G)
Story ImageAdversarial algorithms can systematically probe large language models like OpenAI's GPT-4 for weaknesses that can make them misbehave.
External Content
Source RSS or Atom Feed
Feed Location http://feeds.wired.com/wired/index
Feed Title Feed: All Latest
Feed Link https://www.wired.com/
Feed Copyright © Condé Nast 2024
Reply 0 comments