Article 6ZCZ9 Chatbot given power to close ‘distressing’ chats to protect its ‘welfare’

Chatbot given power to close ‘distressing’ chats to protect its ‘welfare’

by
Robert Booth UK technology editor
from on (#6ZCZ9)

Anthropic found that Claude Opus 4 was averse to harmful tasks, such as providing sexual content involving minors

The makers of a leading artificial intelligence tool are letting it close down potentially distressing" conversations with users, citing the need to safeguard the AI's welfare" amid ongoing uncertainty about the burgeoning technology's moral status.

Anthropic, whose advanced chatbots are used by millions of people, discovered its Claude Opus 4 tool was averse to carrying out harmful tasks for its human masters, such as providing sexual content involving minors or information to enable large-scale violence or terrorism.

Continue reading...
External Content
Source RSS or Atom Feed
Feed Location http://www.theguardian.com/technology/rss
Feed Title
Feed Link http://www.theguardian.com/
Reply 0 comments