OpenAI Stumbled, Allowing Google to Close AI Gap, Says Ex-Research Lead
Ex-OpenAI pioneer argues commercial focus derailed breakthrough research, letting Google seize the AI advantage.
January 22, 2026

The high-stakes race for artificial general intelligence is defined by lightning-fast innovation and the fierce competition between its leading pioneers, making any major shift in the competitive balance a subject of intense scrutiny. That balance has recently been questioned by one of the field's most respected veterans, Jerry Tworek, the former research lead at OpenAI, who asserts that rival Google was able to close the AI capability gap because OpenAI itself stumbled. Tworek, who was instrumental in the foundational work of models like GPT-4 and the advanced reasoning breakthrough of the o1 model series, departed from the company after nearly seven years, lending considerable weight to his critique of the industry leader's trajectory. His comments shine a light on the internal pressures and strategic decisions that are reshaping the frontier of AI development, suggesting that a shift away from pure research in pursuit of rapid commercialization may be costing the company its leading edge.
Tworek’s tenure at OpenAI began in 2019, when the company was still a small team of approximately 30 employees, and his contributions span some of its most transformative projects, including the development of the coding technology behind GitHub Copilot and the post-training of GPT-4.[1][2][3] His most recent and perhaps most critical work centered on a new direction for model intelligence, culminating in the release of the o1 reasoning model. The project, which was known internally by the code names Q-Star and Strawberry, represented a significant conceptual pivot away from the traditional "scaling hypothesis," where intelligence gains primarily relied on ever-increasing amounts of training data and computing power.[4][5] Recognizing the scarcity of new training data and the exponentially rising cost of compute, Tworek's team sought a more "multidimensional" scaling approach, one inspired by human problem-solving.[5] The result was o1, a model trained with reinforcement learning to "think" before generating a response, effectively producing a long, internal "chain of thought."[6][7] This deliberate deliberation process dramatically improved the model's complex reasoning and accuracy, enabling it to achieve Ph.D.-level performance on difficult academic benchmarks in physics, chemistry, and biology, and significantly outperform prior models, including GPT-4o, on numerous challenging reasoning tasks.[6][8] This breakthrough, in Tworek’s view, represented a crucial path forward, providing a more robust and human-like intelligence for future AI systems.
The underlying cause of OpenAI's perceived stumble, according to the analysis surrounding Tworek's high-profile exit, appears to be a growing internal tension between foundational research and commercial demands. Upon his departure, Tworek stated his decision was driven by a desire to "explore types of research that are hard to do at OpenAI," a sentiment widely interpreted as a direct criticism of the company’s increasing and relentless focus on products and revenue.[9][3] Having transformed from a non-profit research lab into a commercial powerhouse with a multi-billion dollar valuation, OpenAI’s operational priorities have shifted toward rapid deployment and market dominance, a change that can often stifle the kind of deep, exploratory, and long-horizon research that yielded innovations like o1.[9] This perceived institutional drift, where fundamental research is increasingly subservient to the product timeline, is the crucial misstep that Tworek’s critique suggests has created an opening for competitors.
While OpenAI was focused on transforming its experimental reasoning model into a commercial product, rival firms have aggressively pursued the same frontier of AI capability. Tworek's concern about a "stumble" is substantiated by the visible advancements from Google, whose flagship Gemini series has quickly become a formidable and direct competitor.[10] Google’s Gemini Ultra model, for instance, claimed to be the first to outperform human experts on the expansive MMLU benchmark, a key measure of knowledge and problem-solving skills across 57 subjects.[10] Furthermore, Google has demonstrated a clear focus on the same core reasoning capabilities pioneered by o1, with a new model effort reportedly aimed directly at rivaling o1's performance in logical reasoning for complex math and programming problems, also utilizing a "chain of thought" approach.[11] This parallel development from a company with vast computational resources and a historical strength in research demonstrates that any delay or deprioritization of breakthrough research at OpenAI is immediately met by a competitive closing of the gap. The AI arms race is defined by instantaneous reaction, and the internal friction over research focus at OpenAI may have created the vital time and space needed for Google and others to level the playing field.
The implications of Tworek’s assessment extend beyond the corporate rivalry of two AI giants; they signal a critical juncture for the entire trajectory of artificial intelligence development. The tension he highlighted—between the necessary patience for non-linear, ambitious research and the immense pressure to monetize and deliver product quickly—is a structural challenge for all frontier AI labs. If the leading institutions fail to maintain an environment where pure, undirected research can flourish, the industry risks an innovation slowdown, where short-term product cycles replace fundamental breakthroughs. Tworek’s work on o1 demonstrated that the path to Artificial General Intelligence (AGI) is not a single, monolithic scaling curve but a quest for new architectural and training paradigms, like teaching a model to "think" with deliberation. The former OpenAI research lead is essentially arguing that by momentarily taking its eye off that crucial, multi-dimensional research ball, the company allowed its nearest and most capable rival to catch up and threaten its position as the undisputed market leader, thereby resetting the competitive landscape to an intense, multi-front war for AI supremacy.[5]