OpenAI leaders ask for regulation of "superintelligent" AI to protect humanity
The leaders suggest other companies working with AI to co-ordinate together to sustain potential risks until a regulatory body is established
OpenAI leaders Wednesday posted a brief note urging for the formation of an international regulator to monitor the use of "superintelligent" AI in order to prevent it from potentially destroying humanity, Guardian reported.
Researchers have already warned of the dangerous risks that AI poses, leading the creators of ChatGPT to call for a regulatory body as big as the International Atomic Energy Agency to contain its risks.
The note posted on the company's website, by the OpenAI co-founders Greg Brockman and Ilya Sutskever and the chief executive, Sam Altman, requested for an international regulator to begin working on how to “inspect systems, require audits, test for compliance with safety standards, [and] place restrictions on degrees of deployment and levels of security." so that the "existential risk" that the technology poses can be sustained.
“It’s conceivable that within the next 10 years, AI systems will exceed expert skill level in most domains, and carry out as much productive activity as one of today’s largest corporations,” they write.
“In terms of both potential upsides and downsides, superintelligence will be more powerful than other technologies humanity has had to contend with in the past. We can have a dramatically more prosperous future; but we have to manage risk to get there. Given the possibility of existential risk, we can’t just be reactive.” they continued.
In the meanwhile, unless a regulatory body is established, the OpenaAI leaders hope for "some degree of co-ordination" among companies working with AI to prioritise safety and care in their projects.
Last week, Altman also expressed worry regarding AI-powered chatbots like ChatGPT saying that they were " a significant area of concern during a congressional hearing in the US Senate.
He also said the technology "required rules and guidelines to prevent misuse." However, he suggested that "introducing licensing and testing requirements for the development of AI" could prevent potential harm from the technology.
-
ISS crew of four completes medical evacuation with safe splashdown off California
-
Annular solar eclipse 2026: Here's everything to know about the ‘ring of fire’
-
World’s first ice archive created to preserve fast-melting glaciers’ secrets
-
NASA, DOE to develop Nuclear Reactor on the Moon by 2030
-
Aurora alert: Northern lights visible tonight at high latitudes
-
Honeybees could help humans communicate with aliens, scientists say
-
Comet 3I/ATLAS: Evidence suggests that interstellar visitor is older than the sun
-
Scientists discovered never-before-seen creature in Great Salt Lake