Analysis-Sam Altman's firing at OpenAI reflects schism over future of AI development

FILE PHOTO: Sam Altman, CEO of Microsoft-backed OpenAI and ChatGPT creator speaks during a talk at Tel Aviv University in Tel Aviv · Reuters

By Greg Bensinger

SAN FRANCISCO (Reuters) - The rift that cost artificial-intelligence whiz kid Sam Altman his CEO job at OpenAI reflects a fundamental difference of opinion over safety, broadly, between two camps developing the world-altering software and pondering its societal impact.

On one side are those, like Altman, who view the rapid development and, especially, public deployment of AI as essential to stress-testing and perfecting the technology. On the other side are those who say the safest path forward is to fully develop and test AI in a laboratory first to ensure it is, so to speak, safe for human consumption.

Altman, 38, was fired on Friday from the company that created the popular ChatGPT chatbot. To many, he was considered the human face of generative AI.

Some caution the hyper-intelligent software could become uncontrollable, leading to catastrophe - a concern among tech workers who follow a social movement called "effective altruism," who believe AI advances should benefit humanity. Among those sharing such fears is OpenAI's Ilya Sutskever, the chief scientist and a board member who approved Altman's ouster.

A similar division has emerged between developers of self-driving cars – also controlled by AI – who say they must be unleashed among dense urban streets to fully understand the vehicles' faculties and foibles; whereas others urge restraint, concerned that the technology presents unknowable risks.

Those worries over generative AI came to a head with the surprise ousting of Altman, who was also OpenAI's cofounder. Generative AI is the term for the software that can spit out coherent content, like essays, computer code and photo-like images, in response to simple prompts. The popularity of OpenAI’s ChatGPT over the past year has accelerated debate about how best to regulate and develop the software.

“The question is whether this is just another product, like social media or cryptocurrency, or whether this is a technology that has the capability to outperform humans and become uncontrollable,” said Connor Leahy, CEO of ConjectureAI and a safety advocate. “Does the future then belong to the machines?”

Sutskever reportedly felt Altman was pushing OpenAI’s software too quickly into users’ hands, potentially compromising safety.

"We don’t have a solution for steering or controlling a potentially superintelligent AI, and preventing it from going rogue," he and a deputy wrote in a July blog post. "Humans won’t be able to reliably supervise AI systems much smarter than us.”