OpenAI co-founder warns ‘superintelligent’ AI must be controlled to prevent possible human extinction

A co-founder of artificial intelligence leader OpenAI is warning that superintelligence must be controlled in order to prevent the extinction of the human race. 

“Superintelligence will be the most impactful technology humanity has ever invented, and could help us solve many of the world’s most important problems. But the vast power of superintelligence could also be very dangerous, and could lead to the disempowerment of humanity or even human extinction,” Ilya Sutskever and head of alignment Jan Leike wrote in a Tuesday blog post, saying they believe such advancements could arrive as soon as this decade. 

They said managing such risks would require new institutions for governance and solving the problem of superintelligence alignment: ensuring AI systems much smarter than humans “follow human intent.” 

“Currently, we don’t have a solution for steering or controlling a potentially superintelligent AI, and preventing it from going rogue. Our current techniques for aligning AI, such as reinforcement learning from human feedback, rely on humans’ ability to supervise AI. But humans won’t be able to reliably supervise AI systems much smarter than us and so our current alignment techniques will not scale to superintelligence,” they wrote. “We need new scientific and technical breakthroughs.”

FBI WARNS OF AI DEEPFAKES BEING USED TO CREATE ‘SEXTORTION’ SCHEMES

OpenAI ChatGPT app on the App Store website

OpenAI ChatGPT app on the App Store website displayed on a screen and OpenAI website displayed on a phone screen are seen in this illustration photo taken in Poland on May 18, 2023. (Jakub Porzycki/NurPhoto)

To solve these problems, within a period of four years, they said they’re leading a new team and dedicating 20% of the compute power secured to date to this effort. 

“While this is an incredibly ambitious goal and we’re not guaranteed to succeed, we are optimistic that a focused, concerted effort can solve this problem,” they said. 

BIDEN ADMIN CLAIMS ‘IRREPARABLE HARM,’ SEEKS TO KEEP MEETING WITH BIG TECH AFTER RULING ORDERED IT TO STOP

Ilya Sutskever

Ilya Sutskever, Russian Israeli-Canadian computer scientist and co-founder and Chief Scientist of OpenAI, speaks at Tel Aviv University in Tel Aviv on June 5, 2023.  (JACK GUEZ/AFP via Getty Images)

In addition to work improving current OpenAI models like ChatGPT and mitigating risks, the new team is focused on the machine learning challenges of aligning superintelligent AI systems with human intent.

Its goal is to devise a roughly human-level automated alignment researcher, using vast amounts of compute to scale it and “iteratively align superintelligence.” 

CLICK HERE TO GET THE FOX NEWS APP 

ChatGPT on a laptop

April 18, 2023, Berlin, Germany: The website of the company OpenAI is open on a laptop.  (Hannes P Albert/picture alliance via Getty Images)

In order to do so, OpenAI will develop a scalable training method, validate the resulting model and then stress test its alignment pipeline. 

The company’s efforts are backed by Microsoft.

Reuters contributed to this report.

Check Also

Larian Studios shocks fans by not planning any Baldur’s Gate 3 DLC or expansions, with no Baldur’s Gate 4 in sight. Time for something new!

During a panel at the Game Developers Conference (GDC) today, the founder of Larian Studios, …