Gary Marcus on the Massive Problems Facing AI & LLM Scaling | The Real Eisman Playbook Episode 42
The Future of AI: Critiques and Insights
Introduction to Gary Marcus
- The discussion opens with a warning about the potential cascading effects on the AI ecosystem if funding for OpenAI diminishes.
- Gary Marcus, a prominent critic of large language models (LLMs), is introduced as a guest who believes that current investments in neural networks may not lead to artificial general intelligence (AGI).
Background of Gary Marcus
- Steve Eisman introduces Gary Marcus, highlighting his critical stance on LLM capabilities, which are central to the AI narrative.
- Marcus shares his lifelong journey studying intelligence and AI, starting from coding at age 10 and focusing on natural intelligence and language acquisition.
Academic Contributions
- His dissertation at MIT explored how children learn language alongside neural networks, emphasizing their marketing appeal rather than true cognitive modeling.
- He reflects on past experiences with neural networks in the '90s, noting their limitations in accurately representing human cognition.
Historical Context of Deep Learning
- In 2012, upon witnessing the resurgence of deep learning, he recognized familiar issues such as hallucinations and reasoning problems that he had previously anticipated.
- He wrote an article for The New Yorker questioning whether deep learning truly represented a revolution in AI.
Key Figures in Deep Learning
- Jeff Hinton is mentioned as a pivotal figure in deep learning who maintained interest during its less popular phases; he won a Nobel Prize recently.
- Hinton's student Ilaskever played a crucial role by demonstrating how GPUs could enhance neural network performance through parallel computation.
Technological Advancements
- The conversation highlights how GPUs were initially designed for video games but became instrumental in running complex neural networks more efficiently.
- This shift allowed researchers to process larger datasets faster than ever before, marking a significant turning point for AI development around 2012.
Understanding Neural Networks and Their Limitations
The Initial Enthusiasm and Critique
- The speaker reflects on writing an article for the New Yorker, praising neural networks while acknowledging their limitations in certain areas.
- Emphasizes that while neural networks excel at pattern recognition and statistical analysis, they struggle with abstraction and reasoning about complex concepts like family trees.
Human Cognition vs. Neural Networks
- References Daniel Kahneman's "Thinking, Fast and Slow," explaining the distinction between System 1 (fast, automatic thinking) and System 2 (slow, deliberative reasoning).
- Argues that neural networks primarily operate like System 1 but lack the capabilities of System 2 reasoning which is essential for human-like intelligence.
Evolution of Large Language Models
- Discusses how the focus has shifted towards large language models (LLMs), a specific type of neural network that emerged after 2012.
- Notes significant investments in LLM development since the introduction of the transformer paper in 2017, estimating trillions spent without careful consideration of cognitive processes.
Naivety in Scaling Intelligence
- Critiques the naive belief that simply scaling up data and computational power would lead to artificial general intelligence.
- Introduces the concept of "trillion-pound baby fallacy," illustrating flawed assumptions about growth rates based on early successes.
Understanding Large Language Models' Functionality
- Explains that LLMs fundamentally predict subsequent elements in a sequence, akin to autocorrect features on smartphones.
- Describes LLMs as "autocomplete on steroids," highlighting their method of breaking down information into smaller parts which can lead to loss of context or connections.
Hallucinations in Language Models
- Points out a critical flaw where LLMs may generate incorrect or fabricated information due to their operational structure.
- Mentions prior warnings about this issue even before LLM technology was fully developed, indicating a long-standing concern regarding their reliability.
Understanding Hallucinations in Language Models
The Nature of Language Models
- Language models break down information into smaller bits to make predictions about what might come next, functioning effectively when trained on vast datasets like the internet.
- They exhibit a form of "glorified memorization," where they can complete text based on previously encountered data, such as finishing paragraphs from well-known works like Harry Potter.
- While they can provide factual answers (e.g., historical sports team locations), they struggle with abstract concepts and may incorrectly reassemble fragmented information.
Defining Hallucinations
- Hallucinations occur when language models generate false information confidently, presenting it as truth without basis in reality.
- An example involves Harry Shearer, who was inaccurately described by a model as a British voiceover actor despite being born in Los Angeles.
Causes of Hallucinations
- A personal anecdote illustrates hallucination: the model claimed the speaker owned a pet chicken named Henrietta, which is entirely fabricated.
- The blending of various pieces of information leads to inaccuracies; for instance, Shearer's fame led to incorrect associations with other British actors.
Implications of Inaccuracies
- Users often mistakenly attribute intelligence to language models; however, they merely reconstruct statistically probable relationships between data points.
- This statistical approach can lead to frequent errors that go unnoticed due to the superficial correctness of generated content.
Real-world Consequences
- Instances have been documented where legal professionals submitted briefs containing fictitious case references generated by AI tools.
- CNET's early use of AI for article writing resulted in numerous errors that editors failed to catch due to the polished appearance of the text.
- The phenomenon termed "work slop" describes reports that look good but contain significant mistakes because LLM outputs lack true understanding.
Understanding Limitations
- Language models do not possess genuine thought processes; they combine elements based on statistical likelihood rather than comprehension or reasoning.
- This inability to discern accuracy means users should remain cautious and critical when interpreting outputs from these systems.
The Impact of LLMs on Society
The Role of LLMs in Information Dissemination
- The speaker expresses skepticism about the reliability of large language models (LLMs), suggesting they present information as if it were factual, regardless of its truthfulness.
- A new article highlights how LLMs are undermining societal institutions, including democracy and civic order, primarily due to their propensity for generating errors.
Consequences for Democracy
- The effectiveness of democracy relies on informed voters; misinformation from LLMs disrupts this process by providing inaccurate data that voters cannot reflect upon meaningfully.
- An example is given regarding the U.S. intervention in Venezuela, where real-time events contradicted what an LLM reported, showcasing a disconnect between reality and AI-generated information.
Limitations of Current AI Models
- The speaker discusses the limitations inherent in AI systems due to their training cutoff dates, which restrict their knowledge to past events and hinder their ability to provide accurate updates.
- A significant issue with these systems is their reliance on memorization rather than understanding context or novelty, leading to failures when faced with unfamiliar scenarios.
Real-world Implications: Tesla Example
- An anecdote about a Tesla car illustrates the dangers of AI systems lacking comprehensive training; the vehicle failed to recognize a jet at an airshow because it had not been trained for such an object.
- This incident underscores the importance of having a general understanding within AI systems to avoid catastrophic mistakes when encountering unexpected situations.
Community Insights and Investment Advice
- Transitioning into investment discussions, Steve Eisman emphasizes the challenge of filtering through overwhelming amounts of financial advice available today.
- He recommends Longle as a community for high-net-worth individuals seeking unbiased investment insights without sales pressure.
Evaluating New AI Models: Gemini vs. ChatGPT
- Discussion shifts towards comparing newer models like Gemini with previous iterations like ChatGPT; improvements are often subjective based on user experience rather than quantifiable metrics.
- Unlike traditional products that can be tested under specific conditions (like cars), evaluating AI performance lacks standardized testing regimes, complicating assessments.
Understanding the Evolution of Language Models
The Impact of Early Errors in Computing
- Discussion on a historical error in Intel's floating point calculations, highlighting that even minor inaccuracies can lead to significant scrutiny and debate.
- Emphasis on the subjective nature of evaluating language models, as different users have varying needs and experiences with each model.
Comparing Language Models: A Subjective Experience
- Recognition that comparisons between models like Gemini and GPT-5.2 depend heavily on individual use cases, such as coding or brainstorming.
- The speaker expresses a desire for innovative models that demonstrate unique capabilities rather than incremental improvements.
Diminishing Returns in Model Development
- Explanation of diminishing returns observed in newer models; while advancements continue, they are less dramatic compared to earlier versions.
- Confidence in assessing model improvements based on benchmarks and intuitive testing, noting how early iterations were clearly inferior.
User Experience vs. Benchmarking
- Personal anecdotes about the clear superiority of GPT-2 over GPT-1 without needing formal tests; similar observations made for subsequent versions up to GPT-4.
- Introduction of subtlety in comparing GPT-5 with its predecessor, indicating a need for more structured measures to evaluate progress.
Business Implications and GPU Investments
- Discussion on the massive investments by hyperscalers like Microsoft into GPU technology for LLM development, estimated at around $500 billion.
- Uncertainty surrounding the exact percentage of Nvidia GPUs allocated to LLM companies but acknowledgment that it is likely substantial.
Shifts in Community Perspectives
- Speculation about potential shifts within tech companies if new perspectives (like those from Gary Marcus) gain traction regarding LLM development strategies.
- Reflection on changes within the AI research community over recent months, suggesting a growing consensus among long-time researchers against mainstream narratives.
Survey Insights from AI Researchers
- Reference to a survey conducted by AAAI revealing that 85% of respondents (mostly academics and industry researchers) may not align with popular media portrayals or corporate views on AI advancements.
The Limitations of Large Language Models and the Shift in AI Perspectives
The Reality of Artificial General Intelligence
- A significant 85% of companies believe that large language models (LLMs) will not lead to achieving artificial general intelligence (AGI), indicating a growing skepticism within the industry.
- The speaker has been vocal about these limitations for a long time, suggesting that many others are now beginning to recognize similar concerns regarding LLMs and AGI.
- Ilya Sutskever, co-founder of OpenAI, recently expressed doubts about the effectiveness of current AI models, marking a notable shift in discourse from key figures in the field.
Turning Points in AI Development
- The turning point for public perception occurred around August 2023 when Sam Altman hinted at GPT-5 being close to AGI, which raised expectations significantly.
- Anticipation for GPT-5 was likened to waiting for an event with repeated delays, leading to frustration among users who expected groundbreaking advancements.
Public Disappointment with GPT-5
- When GPT-5 was finally released on August 7th, 2025, it failed to meet inflated expectations; users quickly recognized its limitations despite some improvements over previous versions.
- The community's disappointment led to discussions on social media acknowledging that critiques made by Gary Marcus were valid and resonated widely after the release.
Changes in Industry Practices
- There is a noticeable shift within companies towards integrating classical symbolic AI alongside LLMs as they begin recognizing their limitations.
- Companies are adopting code interpreters and other symbolic tools quietly but effectively improving system performance without making it publicly known.
Insights into Future Directions
- The integration of symbolic tools is crucial as they operate differently than traditional LLM frameworks, hinting at a potential evolution in how AI systems are developed moving forward.
The Implications of AGI Development and Market Dynamics
Departure from Major Tech Companies
- The trend of employees leaving major tech companies like OpenAI to start their own startups suggests a lack of confidence in the imminent release of AGI, as one would expect them to stay for such a significant event.
- The mass exodus indicates that insiders may not believe in the revolutionary capabilities they are purportedly developing, hinting at potential overestimations regarding AGI readiness.
Competitive Landscape in AI
- The argument is made that scaling large language models (LLMs) has become a race where only those with substantial resources, like Google, can prevail due to the absence of unique technological advantages among competitors.
- A price war has emerged as LLM technology becomes commoditized; prices have dropped significantly, which ultimately favors larger players like Google who can absorb these costs better than smaller firms.
- As all companies follow similar development paths without any "secret sauce," there is an expected convergence at the top tier of AI models, leading to minimal differentiation between offerings.
Inference Models vs. LLM Models
- Inference models operate on LLM frameworks but differ by taking multiple passes to refine answers rather than providing immediate responses. This iterative process aims for higher accuracy and quality in outputs.
- Traditional neural networks typically provide quick responses through single-pass processing; however, new inference paradigms allow for variable computation times based on problem complexity.
Applications and Limitations of Inference Models
- Inference models excel in domains with verifiable data generation capabilities, such as mathematics or programming, where structured problems allow for effective training and solution generation.
- They struggle with open-ended real-world scenarios due to their reliance on closed datasets; unexpected variables can lead to failures similar to past financial model collapses when faced with unforeseen market conditions.
Understanding the Limitations of AI and the Path Forward
The Challenge of Novelty in AI
- AI systems perform well when operating within familiar parameters but struggle with novel situations, as illustrated by the Tesla example. This highlights a fundamental issue with novelty in AI.
- Most interesting real-world scenarios involve some level of novelty, which current models are not equipped to handle effectively, especially outside narrow domains like chess or Go.
- In complex fields such as politics or military strategy, historical data alone is insufficient for reasoning about unprecedented events; new concepts must be considered.
- Addressing unique situations requires understanding abstract concepts like power and diplomacy rather than relying solely on past data.
- Even in customer service, where interactions seem straightforward, variations in user queries can lead to system failures due to their inability to adapt to novelty.
Recommendations for Advancing AI Development
- If appointed as an advisor (AI ZAR), one would emphasize the need for greater intellectual diversity among teams working on AI technologies.
- The field has overly focused on scaling large language models (LLMs), which has proven inefficient and costly without achieving true artificial general intelligence (AGI).
- Current systems require vast amounts of data compared to human learning processes, indicating inefficiencies that need addressing through alternative approaches.
- Investment strategies should shift towards more reliable and economical methods rather than continuing down a path that prioritizes high-cost solutions yielding limited results.
- Venture capitalists often prioritize plausible-sounding ideas over effective solutions due to financial incentives, leading to wasted resources in the pursuit of scaling LLM technology.
Market Dynamics and Speculation
- While some venture capitalists genuinely aim to advance technology, many operate cynically by investing in expensive ideas that yield substantial cuts regardless of their success.
- This focus on scaling has not been intellectually sound nor beneficial overall; it risks significant losses for investors while providing short-term gains for VCs.
- The market's irrationality can persist longer than expected; however, signs indicate growing skepticism within investment circles regarding returns from current AI investments.
- Investors are beginning to recognize issues such as circular financing and poor ROI from existing systems since November 2022, prompting reevaluation of strategies.
- Despite Nvidia's impressive products and ecosystem, speculation around infinite demand raises concerns about sustainability if AGI does not materialize as anticipated.
The Vulnerability of OpenAI in the AI Market
Current State of AI Job Integration
- A recent study reported that only 2.5% of jobs performed with AI systems are actually feasible, indicating a gap between expectations and reality regarding AI capabilities.
- This discrepancy suggests that significant investments in AI technology, particularly chips, may not be justified.
OpenAI's Financial Challenges
- OpenAI is seen as vulnerable due to its substantial financial commitments without profitability; it faces intense competition from Google, which has surpassed it technologically.
- Concerns arise about OpenAI being compared to WeWork in terms of valuation; despite billions in revenue, they incur massive monthly losses.
Competitive Landscape
- Google's advancements position it favorably against OpenAI; if success relies solely on scale, Google is likely to dominate due to its resources and infrastructure.
- The need for large funding amounts (potentially $100 billion) raises concerns about OpenAI's sustainability if major investors withdraw support.
Future Prospects for OpenAI
- If funding becomes scarce, Microsoft may absorb OpenAI as a potential solution to its financial struggles.
Understanding World Models in AI
- The discussion shifts towards the concept of "world models," essential for effective artificial intelligence systems.
- A world model represents external realities within software; it's crucial for applications like GPS navigation systems to function accurately.
Historical Context and Importance of World Models
- Historically, world models have been foundational in classical AI development; notable figures like Herb Simon emphasized their necessity for effective problem-solving.
Limitations of Current Language Models
- Large language models often lack proper world models leading to inaccuracies (e.g., incorrect facts), highlighting the need for structured knowledge representation.
Human vs. Machine Understanding
- Humans naturally create mental models from narratives (like movies or books), allowing them to discern plausibility within fictional worlds—something current AI struggles with.
Understanding AI's Limitations and Future
The Challenge of Classical AI
- Classical AI excels at manual problem-solving, as demonstrated by Doug Lenat's work on "Romeo and Juliet," where his system could understand key plot points without relying on secondary sources.
- Despite its capabilities, classical AI struggles to acquire models autonomously; large language models often simulate understanding rather than genuinely grasping concepts.
Limitations of Current Models
- An example highlighting this limitation is chess: even after training on extensive data, models can still make illegal moves, indicating a failure to abstract the underlying principles of the game.
- There is a pressing need for systems that can induce world models from data, understanding causal relationships and entities involved—a complex challenge not easily resolved.
The Path Forward in AI Research
- Transformative potential exists within artificial intelligence, but current technologies are insufficient for immediate breakthroughs; foundational research is essential for future advancements.
- Investments should focus on genuine research rather than speculation about scaling existing technologies. Simply increasing size will not address fundamental issues in AI development.
Conclusion and Reflection
- The discussion emphasizes the importance of listening to experts advocating for foundational research in AI. Acknowledging these insights could lead to more effective strategies moving forward.