OpenAI leaders call for regulation to prevent AI destroying humanity
Team behind ChatGPT say equivalent of atomic watchdog is needed to guard against risks of superintelligent' AIs
The leaders of the ChatGPT developer OpenAI have called for the regulation of superintelligent" AIs, arguing that an equivalent to the International Atomic Energy Agency is needed to protect humanity from the risk of accidentally creating something with the power to destroy it.
In a short note published to the company's website, co-founders Greg Brockman and Ilya Sutskever and the chief executive, Sam Altman, call for an international regulator to begin working on how to inspect systems, require audits, test for compliance with safety standards, [and] place restrictions on degrees of deployment and levels of security" in order to reduce the existential risk" such systems could pose.
Continue reading...