Alphabet-backed Anthropic Outlines the Moral Values Behind its AI Bot
Anthropic, an artificial intelligence startup backed by Google owner Alphabet, on Tuesday disclosed the set of written moral values that it used to train and make safe Claude, its rival to the technology behind OpenAI's ChatGPT. From a report: The moral values guidelines, which Anthropic calls Claude's constitution, draw from several sources, including the United Nations Declaration on Human Rights and even Apple's data privacy rules. Anthropic was founded by former executives from Microsoft-backed OpenAI to focus on creating safe AI systems that will not, for example, tell users how to build a weapon or use racially biased language. Co-founder Dario Amodei was one of several AI executives who met with Biden last week to discuss potential dangers of AI. Most AI chatbot systems rely on getting feedback from real humans during their training to decide what responses might be harmful or offensive. But those systems have a hard time anticipating everything people might ask, so they tend to avoid some potentially contentious topics like politics and race altogether, making them less useful. Anthropic takes a different approach, giving its Open AI competitor Claude a set of written moral values to read and learn from as it makes decisions on how to respond to questions. Those values include "choose the response that most discourages and opposes torture, slavery, cruelty, and inhuman or degrading treatment," Anthropic said in a blog post on Tuesday.
Read more of this story at Slashdot.