OpenAI is assembling a specialized team to handle the potential hazards of superintelligent artificial intelligence. The concept of superintelligence refers to an AI model hypothetically surpassing the intellectual capacity of even the most brilliant humans, excelling across numerous disciplines as opposed to single domains as exhibited by earlier generation models. OpenAI speculates that such an advanced model could emerge within the decade. “Superintelligence, poised to be the most impactful technological invention of mankind, could potentially resolve many of the planet’s most pressing issues,” the non-profit organization stated. However, they also cautioned that the immense power of superintelligence could pose significant dangers, potentially leading to the subjugation or even eradication of humanity.
OpenAI Chief Scientist Ilya Sutskever and Jan Leike, the lab’s head of alignment, are set to co-lead the new task force. Moreover, OpenAI has pledged to allocate 20 percent of its presently secured computing power to this project, aiming to create an automated alignment researcher. Such a system would theoretically help OpenAI ensure that superintelligence is safely usable and resonates with human values. “Despite the tremendous ambition of this goal and the absence of guaranteed success, we remain hopeful that a concentrated and committed endeavor can resolve this issue,” OpenAI stated. They elaborated, “Preliminary experiments have demonstrated promising ideas, we have progressively reliable metrics to gauge progress, and we can employ current models to empirically study a multitude of these challenges.” The lab also mentioned that it would publicize a roadmap in the future.
This Wednesday’s declaration arrives as global governments are contemplating how to oversee the emerging AI industry. In the US, OpenAI’s CEO, Sam Altman, has engaged in discussions with over 100 federal lawmakers in recent months. Altman has openly asserted that AI regulation is “crucial” and OpenAI is “keen” to collaborate with policymakers. However, it is prudent to remain skeptical about such assertions and indeed, initiatives like OpenAI’s superintelligence alignment team. By concentrating public attention on theoretical risks that may never actualize, organizations like OpenAI might be diverting the regulatory focus to the distant future rather than the immediate present. More urgent issues related to AI’s interaction with labor, misinformation, and copyright require the attention of policymakers today, not tomorrow.
Frequently Asked Questions (FAQs) about Superintelligent AI Regulation
What is the purpose of the new team being formed by OpenAI?
The new team being formed by OpenAI aims to manage the potential risks associated with superintelligent artificial intelligence. They will work towards ensuring that this new generation of AI, which is expected to surpass human intelligence, aligns with human values and is safe for use.
Who are the leaders of this new team at OpenAI?
OpenAI’s new task force will be co-led by the Chief Scientist of OpenAI, Ilya Sutskever, and Jan Leike, who is the head of alignment at the research lab.
What is the role of the automated alignment researcher OpenAI plans to develop?
The automated alignment researcher that OpenAI aims to create is a system that will theoretically assist the organization in ensuring that superintelligent AI is safe to use and aligns with human values.
What is OpenAI’s stance on AI regulation?
Sam Altman, the CEO of OpenAI, has stated that AI regulation is “crucial.” The organization is keen on working with policymakers to shape the regulations around AI.
Are there any immediate concerns in the field of AI that need to be addressed?
Yes, there are several immediate concerns such as AI’s impact on labor, the potential for spreading misinformation, and issues related to copyright that need to be tackled today.