Article 70N73 It's trivially easy to poison LLMs into spitting out gibberish, says Anthropic

It's trivially easy to poison LLMs into spitting out gibberish, says Anthropic

by
from The Register on (#70N73)
Story ImageJust 250 malicious training documents can poison a 13B parameter model - that's 0.00016% of a whole dataset

Poisoning AI models might be way easier than previously thought if an Anthropic study is anything to go on....

External Content
Source RSS or Atom Feed
Feed Location http://www.theregister.co.uk/headlines.atom
Feed Title The Register
Feed Link https://www.theregister.com/
Feed Copyright Copyright © 2025, Situation Publishing
Reply 0 comments