OpenAI Co-founder Andrej Karpathy Questions AI Hype: ‘The Models Are Not There’

Andrej Karpathy’s Sobering Assessment Fuels Debate Over the AI Bubble

Andrej Karpathy, a founding member of OpenAI and one of the most respected voices in deep learning, has delivered a stark assessment of the current state of large language models (LLMs), suggesting that the technology powering the recent AI boom is fundamentally limited. His comments, made during a recent interview on The Dwarkesh Podcast, have injected a dose of skepticism into a market currently defined by massive valuations and intense venture capital investment.

Karpathy, who previously served as the Director of AI at Tesla before returning to OpenAI, stated plainly that, despite their impressive capabilities, “the models are not there” yet. This critique comes at a critical time in 2025, when the AI sector is valued in the trillions and companies like Nvidia have seen unprecedented stock surges based on the promise of generative AI.


The Core Critique: Why Current LLMs Fall Short

Karpathy’s analysis centers on the inherent limitations of the current generation of LLMs, which he characterizes as sophisticated pattern-matchers rather than true reasoning engines. He argues that the industry needs to move beyond the current paradigm of simple text prediction.

Next Token Predictors, Not Reasoning Engines

According to Karpathy, the current models, including those developed by OpenAI, are essentially “next token predictors.” This means their primary function is to statistically determine the most plausible next word or token in a sequence, based on the vast data they were trained on. While this process results in fluent, coherent, and often astonishingly human-like text, it masks a lack of genuine understanding or planning capability.

Karpathy referenced the critical term “stochastic parrots,” a phrase used by AI researchers to describe models that excel at mimicking language patterns without internalizing meaning or context. He highlighted several key areas where current LLMs fail to meet the expectations set by market hype:

  • Lack of Reliability: Models frequently hallucinate or generate factually incorrect information because they prioritize statistical plausibility over truth.
  • Poor Multi-Step Reasoning: They struggle with complex tasks that require sustained, multi-step logical deduction or planning, often failing early in the sequence.
  • Absence of Memory and State: LLMs lack a robust, long-term memory or the ability to maintain a consistent internal state across extended interactions, making them poor agents for complex, ongoing tasks.

“The models are not there. They are not reliable. They are not reasoning engines. They are next token predictors that are very good at pattern matching,” Karpathy stated, underscoring the gap between current technology and the vision of truly intelligent AI.


The AI Bubble Question and Market Implications

Karpathy’s comments are significant because they come from an insider who has been instrumental in building the technology driving the current boom. His skepticism directly challenges the narrative that current LLMs represent the final, scalable architecture for Artificial General Intelligence (AGI).

The Disconnect Between Hype and Reality

For investors and business leaders, Karpathy’s critique serves as a warning sign that the massive capital pouring into AI might be based on an overestimation of the technology’s immediate capabilities. The current market environment is characterized by:

  • Sky-High Valuations: AI startups are commanding valuations in the tens of billions based on future potential, often before demonstrating sustainable profitability.
  • Infrastructure Spending: Billions are being spent on specialized hardware (like Nvidia GPUs) to train and run these models, assuming exponential performance gains will continue indefinitely.

If the fundamental architecture of LLMs is hitting a wall—as Karpathy suggests—then the returns on these investments may be slower or less transformative than anticipated, potentially leading to a correction in the market often referred to as the “AI bubble.”

Historical Context: The AI Winter Precedent

This is not the first time the AI field has faced a reality check. The industry has experienced several “AI Winters”—periods of reduced funding and interest following overhyped expectations. Experts like Karpathy are keen to avoid a similar cycle by setting realistic expectations for current capabilities and focusing resources on genuine architectural breakthroughs, rather than simply scaling up existing models.


The Path to True Reasoning: What Comes Next

Karpathy did not merely criticize; he also outlined the necessary steps for the next major leap in AI development. He suggests that the future lies in models that move beyond mere statistical prediction and incorporate true reasoning and planning capabilities.

Key Requirements for Next-Generation AI

To achieve reliable, intelligent systems, Karpathy believes researchers must focus on integrating mechanisms that allow models to:

  1. Maintain State and Memory: Develop architectures that allow the model to remember and utilize information from long-past interactions, maintaining a consistent internal world model.
  2. Execute Planning and Search: Equip models with the ability to look ahead, evaluate multiple potential outcomes, and plan multi-step actions, similar to how humans solve complex problems.
  3. Improve Reliability: Introduce mechanisms that inherently prioritize factual accuracy and logical consistency over simply generating statistically plausible text.

This shift implies a move away from purely transformer-based LLMs toward more complex, hybrid architectures that incorporate symbolic reasoning or other methods to ground the models in reality and logic.


Key Takeaways for Business and Technology Leaders

Karpathy’s intervention offers crucial guidance for companies currently integrating or investing heavily in generative AI:

  • Manage Expectations: Current LLMs are powerful tools for content generation and summarization, but they should not be trusted with mission-critical tasks requiring absolute factual accuracy or complex, unverified reasoning.
  • Focus on Specific Applications: The highest value currently lies in narrow, well-defined applications where the model’s pattern-matching strengths can be leveraged (e.g., coding assistance, data synthesis, creative drafting).
  • Anticipate Architectural Shifts: Investors and developers should prepare for the possibility that the next generation of AI may require fundamentally different hardware and software architectures than the current LLM paradigm.

Conclusion

Andrej Karpathy’s candid assessment serves as a necessary reality check for the overheated AI market. While the impressive progress of the last few years is undeniable, his warning that current LLMs are still fundamentally limited “stochastic parrots” highlights the significant technological hurdles that remain before true Artificial General Intelligence is realized. For the business world, this means tempering the hype with practical application and recognizing that the current AI boom is still in its early, experimental phase, requiring continued, focused research into foundational architectural improvements.

Source: Fortune

Original author: Eva Roytburg

Originally published: October 21, 2025

Editorial note: Our team reviewed and enhanced this coverage with AI-assisted tools and human editing to add helpful context while preserving verified facts and quotations from the original source.

We encourage you to consult the publisher above for the complete report and to reach out if you spot inaccuracies or compliance concerns.

Author

  • Eduardo Silva is a Full-Stack Developer and SEO Specialist with over a decade of experience. He specializes in PHP, WordPress, and Python. He holds a degree in Advertising and Propaganda and certifications in English and Cinema, blending technical skill with creative insight.

Share this: