OpenAI opens the door for military uses but maintains AI weapons ban
Enlarge (credit: OpenAI / Getty Images / Benj Edwards)
On Tuesday, ChatGPT developer OpenAI revealed that it is collaborating with the United States Defense Department on cybersecurity projects and exploring ways to prevent veteran suicide, reports Bloomberg. OpenAI revealed the collaboration during an interview with the news outlet at the World Economic Forum in Davos. The AI company recently modified its policies, allowing for certain military applications of its technology, while maintaining prohibitions against using it to develop weapons.
According to Anna Makanju, OpenAI's vice president of global affairs, "many people thought that [a previous blanket prohibition on military applications] would prohibit many of these use cases, which people think are very much aligned with what we want to see in the world." OpenAI removed terms from its service agreement that previously blocked AI use in "military and warfare" situations, but the company still upholds a ban on its technology being used to develop weapons or to cause harm or property damage.
Under the "Universal Policies" section of OpenAI's Usage Policies document, section 2 says, "Don't use our service to harm yourself or others." The prohibition includes using its AI products to "develop or use weapons." Changes to the terms that removed the "military and warfare" prohibitions appear to have been made by OpenAI on January 10.