OpenAI, the leading player in the pursuit of Artificial General Intelligence (AGI), found itself embroiled in controversy with the sudden dismissal of CEO Sam Altman. Despite Altman’s reinstatement, reports indicate that his initial removal stemmed from board disagreements over AGI safeguards, honesty concerns, and worries about hastening the release of AI products to market without sufficient safety precautions.

Speculation is rife regarding OpenAI’s AGI progress, with unnamed sources suggesting that researchers may be closer to achieving AGI than publicly disclosed. The board’s worry over Altman’s alleged haste in capitalizing on ChatGPT’s success for new AI products, possibly linked to AGI, adds a layer of complexity to the unfolding drama.

The definition of AGI itself has evolved within OpenAI, from tasks teachable to humans to surpassing human capabilities in economically valuable domains. Transparency has reportedly declined since the release of GPT-2 in 2019, with an increase in secrecy surrounding model details, potentially tied to AGI development.

The implications of AGI are profound, requiring global collaboration to address risks and mitigation efforts. Concerns loom over job displacement and social upheaval, with the potential concentration of power in the hands of the wealthy once AGI arrives.

When Altman was fired, the majority of OpenAI’s 700 employees threatened to quit unless existing board members resigned, and Altman was reinstated. Before Altman’s dismissal, staff researchers sent a warning letter to the board about a powerful AI discovery, potentially linked to the Q* project. 

Q* is seen as a breakthrough in the quest for AGI, showcasing remarkable mathematical problem-solving abilities with significant computing resources. The letter flagged the prowess and potential dangers of AI without specifying safety concerns.

The Q* project and the warning letter were cited as contributing factors in Altman’s firing, with Mira Murati, a long-time executive, mentioning the project to employees. An OpenAI spokesperson confirmed Murati’s communication but did not comment on the accuracy of the reports.

Altman’s leadership in propelling ChatGPT’s success and his tantalizing tease about AGI being in sight at a gathering of world leaders in San Francisco add context to the significance of the Q* project. The board’s decision to terminate Altman, strategically timed a day after his public mention of AGI’s progress, has sent shockwaves through the tech industry, marking an unprecedented news cycle with potential links to AGI safety concerns.

As the OpenAI saga continues to unfold, the tech world braces for the impact of this dramatic turn of events, pondering the future of AGI and the organization at the forefront of its development.

Related: