Some anonymous whistleblower suggests OpenAI initiated the development of a 125 trillion parameter multimodal model named Rakus, or QSTAR, in August 2022, which completed its training phase by December 2023. However, the launch was postponed owing to the prohibitive costs associated with inference. The narrative further unfolds to reveal a series of planned developments leading to AGI, including the renaming and cancellation of various model releases in response to evolving challenges and strategic decisions within OpenAI.
Amidst these revelations, the video references a lawsuit by Elon Musk, accusing OpenAI of deviating from its foundational mission to democratize access to advanced AI technologies. This legal action, according to the whistleblower, has influenced the timeline of AGI development, underscoring the intricate web of technical, legal, and ethical considerations that shape the trajectory of AI advancements.
The video doesn’t just speculate on future models and corporate maneuvers; it also engages with the broader implications of AGI. The whistleblower reflects on the different levels of AGI, from emerging capabilities that match or exceed those of an unskilled human, to virtuoso and superintelligent systems that far surpass human expertise across all domains. This nuanced discussion highlights the complexity of AGI development, suggesting that achieving AGI involves not only quantitative leaps in computational power but also qualitative advancements in AI’s cognitive and adaptive capabilities.

A significant portion of the discussion is dedicated to the exponential increase in the parameter count of AI models, drawing parallels between the neural networks of AI systems and the synaptic connections of biological brains. The video posits that the sheer scale of parameters could be a critical factor in bridging the gap between narrow AI applications and the multifaceted adaptability required for AGI.
The whistleblower’s analysis is peppered with cautionary notes about the speculative nature of the document and the predictions derived from it. Yet, they argue that the detailed planning and incremental advancements outlined in the document, when viewed alongside OpenAI’s public endeavors and the broader trajectory of AI research, lend a degree of credibility to the speculated timeline for achieving AGI by 2027.
In essence, the video serves as a fascinating, if speculative, glimpse into the potential future of AI development, marked by groundbreaking technological leaps, strategic pivots, and the looming challenge of navigating the ethical and societal implications of creating machines that could one day match or exceed human intelligence across all domains. The whistleblower’s insights underscore the importance of transparency, ethical considerations, and the need for a collaborative approach to governance as humanity stands on the brink of potentially realizing AGI.