This New AI Model Thinks Without Language (w/ Eve Bodnia of Logical Intelligence)

This New AI Model Thinks Without Language (w/ Eve Bodnia of Logical Intelligence)

Understanding Energy-Based Models and AGI

Introduction to Abstract Thinking in AI

  • The speaker emphasizes that intelligence is not solely language-based, suggesting a more abstract form of thinking.
  • They introduce the concept of energy-based models (EBM), which differ from traditional language models by eliminating the guessing game associated with predicting the next word.

Major Announcement from Logical Intelligence

  • The podcast discusses significant news regarding Logical Intelligence, highlighting Yan Lun's appointment as founding chair of their technical research board.
  • Eve Bodnia, founder of Logical Intelligence, has an impressive background in physics and claims her new Kona model shows credible signs of artificial general intelligence (AGI).

Exploring Energy-Based Models

  • The discussion will focus on what energy-based models are, their advantages over language models, and their potential role in achieving AGI versus merely functioning as strong constraint solvers.

Eve Bodnia's Journey to Founding Logical Intelligence

  • Eve shares her complex journey into logical intelligence, driven by a childhood curiosity about the universe and a desire for limitless exploration in theoretical science.
  • She reflects on her academic path, choosing physics due to its foundational nature and her aptitude for mathematics.

Academic Influences and Collaborations

  • Eve describes how exposure to various scientific areas led her to recognize patterns applicable both in particle physics and understanding brain functions.
  • Her collaboration with Michael Freriedman at Google Quantum AI sparked discussions about fundamental laws governing intelligence from a physics perspective rather than traditional computer science methods.

Transitioning from Academia to Industry

  • Initially focused on academia, Eve was encouraged by peers to consider starting a company due to the rapid pace of advancements in AI compared to academic settings.
  • Despite initial reluctance while being pregnant and settled at home, she ultimately embraced the idea of launching a tech venture.

Understanding Energy-Based Models vs. Language Models

The Importance of Profitability and Vision Alignment

  • The speaker emphasizes the necessity of profitability in business while balancing various factors, highlighting the importance of having aligned vision with trusted partners and investors.
  • They express satisfaction with their decision to scale operations, indicating that they are still engaged in their original work but at a larger scale.

Overview of Energy-Based Models (EBMs)

  • A request is made for a simple overview of energy-based models, particularly in contrast to language models.
  • The speaker reflects on their initial experiences with large language models (LLMs), noting their effectiveness for language-related tasks but also recognizing limitations such as inaccuracies when handling complex queries.

Limitations of Language Models

  • The speaker discusses how LLMs operate by mapping data into a linguistic space and predicting subsequent words, which can lead to "hallucinations" or inaccuracies due to the nature of language proximity.
  • They share personal insights about thinking abstractly across multiple languages, suggesting that intelligence isn't strictly tied to any single language.

Abstract Thinking Beyond Language

  • The realization emerges that not all information is linguistically bound; robotics and other forms of intelligence can function without relying on language.
  • The speaker argues that many people mistakenly equate AI solely with LLM capabilities, emphasizing the existence of alternative AI models capable of abstract thought.

Exploring Non-Linguistic Communication

  • They illustrate that communication can occur through various forms beyond text or speech—such as visual art or music—highlighting the need for diverse methods in AI interaction.
  • An example is provided using Sudoku puzzles, demonstrating problem-solving without linguistic patterns, reinforcing the idea that spatial thinking plays a crucial role in cognition.

Characteristics of Energy-Based Models

  • EBMs are described as operating independently from traditional linguistic frameworks; they utilize an abstract representation akin to machine language rather than tokens.
  • This model allows flexibility in output formats—images, videos, or direct software interactions—without being constrained by tokenization processes typical in LLM applications.

Navigating Abstract Representations

  • The discussion concludes with an emphasis on navigating an "energy landscape," where scenarios are assessed simultaneously without reliance on tokens.
  • Clarification is provided regarding voice models being considered a form of language; however, EBMs offer unique pathways for processing data outside conventional linguistic structures.

Understanding AGI and Its Implications

The Nature of Language and Intelligence

  • The speaker discusses how language is a manifestation of intelligence, emphasizing that while one can mimic language, true understanding requires cognitive attachment.
  • They introduce the concept of an ecosystem around Artificial General Intelligence (AGI), noting that definitions of AGI vary among individuals.

Defining AGI

  • The speaker stresses the importance of defining AGI, suggesting it encompasses general intelligence capable of planning and adaptation.
  • They explain that human intelligence involves memory systems that aid in planning and prediction, which are essential for survival.

Adaptation and Response to Environment

  • A key aspect of intelligence is the ability to respond quickly to changing environments, such as unpredictable weather or manufacturing situations.
  • The discussion highlights that effective adaptation is crucial for safety in AI applications like self-driving cars.

Evolutionary Perspective on AI

  • The speaker proposes a definition of AGI based on its ability to evolve by adapting intelligently to tasks while preserving objectives.
  • They clarify that evolution in AI should aim at minimizing resource use while achieving specific goals effectively.

Concerns About AI Implementation

  • There are concerns regarding self-driving cars using language models due to potential unpredictability in decision-making processes.
  • Emphasizing safety, the speaker notes the need for reliable AI systems as they become ubiquitous in society over the next five years.

The Role of Hallucinations in Intelligence

Understanding Human-Like Hallucinations

  • The conversation touches on natural hallucinations experienced by humans and their implications for building precise engineering solutions.

Mechanisms for Self-Aligned Models

  • Discussion shifts towards how AI can constrain itself through training methods that map input data within an energy landscape framework.

Energy Landscape Concept

  • The speaker explains how engineers can set constraints during model training by mapping data onto an energy landscape where lower points represent more probable scenarios.

Application of Physics Principles

  • They draw parallels between theoretical physics modeling and AI training, focusing on minimizing energy states to predict model behavior effectively.

This structured approach provides a comprehensive overview while allowing easy navigation through timestamps linked directly to relevant sections.

Understanding Energy-Based Models and Their Applications

Perturbation Theory in Modeling

  • The discussion begins with the concept of perturbation theory, which allows models to return to their original landscape despite minor adjustments. This is crucial for understanding how different modeling approaches can affect outcomes.
  • The speaker emphasizes that large language models (LLMs) struggle with self-alignment due to architectural differences, making it difficult to apply perturbation theory effectively.

Hallucination-Free Tasks and Verification

  • For tasks requiring precise answers, such as data analysis or code generation, external verifiers like Lean 4 can be used to ensure correctness. This highlights the importance of formal verification in computational tasks.
  • The speaker mentions a Sudoku test on their website, showcasing the speed and efficiency of their model compared to others.

Energy-Based Models vs. Traditional Approaches

  • The energy-based model eliminates guessing by providing a clear energy landscape where the correct answer is identifiable quickly, thus saving time during computations.
  • The speaker notes that their models are relatively small yet efficient, running on affordable hardware while achieving significant performance improvements.

Brain Functionality and Model Efficiency

  • A comparison is made between human brain efficiency (operating under 20 watts) and machine requirements for similar tasks. This raises questions about optimizing architectures for better performance without excessive resource consumption.

Diffusion Models and Energy-Based Reasoning

  • There’s an exploration of similarities between diffusion models and energy-based reasoning models; however, clarity in definitions is emphasized to avoid ambiguity in discussions.
  • While energy-based techniques have been around for decades, the novelty lies in designing reasoning parts within these models rather than merely applying existing methods.

Limitations of LLMs in Extrapolation

  • The conversation shifts towards LLM capabilities; they are often perceived as intelligent but lack true extrapolation abilities across different domains—an essential trait of natural intelligence.
  • Unlike children who can transfer knowledge across various fields after learning specific skills (like mathematics), LLMs remain limited to what they were specifically trained on.

This structured overview captures key insights from the transcript while linking back to specific timestamps for further exploration.

Development of Kona: Current Status and Future Roadmap

Overview of Development Progress

  • The speaker expresses skepticism about the feasibility of certain developments, indicating that they believe some goals may never be achieved.
  • Initially, the company started with a theoretical idea which evolved into a proof of concept thanks to talented engineers. This transition occurred over several months.
  • After designing the architecture, experiments were conducted to assess compatibility with large language models (LLMs) and transformers due to their fundamental differences.

Experimentation and Scaling

  • The team successfully scaled down from a transformer model to a simpler version related to LLMs, confirming its functionality through various tests.
  • They attached an actual LLM to the energy-based model (EBM), testing its effectiveness as a user interface for prompting EBM tasks.

Benchmarking and Challenges

  • A series of benchmarks were established for different model versions, comparing performance across smallest versions, proof of concepts, and real models.
  • Despite having a solid theoretical understanding, engineering challenges arose during integration due to the non-autoregressive nature of EBM compared to autoregressive transformers.

Information Loss Issues

  • Significant information loss was encountered when integrating LLM prompts with EBM due to differing operational methodologies between them.
  • The architecture is now scalable and progressing beyond initial expectations after overcoming major difficulties in orchestration.

Future Directions and Applications

  • The speaker discusses whether LLM attachment will always be necessary or if it’s just an interim step; they emphasize the importance of adaptability in AI systems for human interaction.
  • While not essential for all applications (like robotics), having language capabilities can enhance functionalities such as spatial navigation or data analysis in real-time scenarios.

Integration with Sensory Data

  • There is potential for integrating sensory data (e.g., cameras or temperature sensors), allowing models like EBMs to operate without separate video/image generation models.
  • This integration could streamline processes by mapping various sensory inputs into one cohesive space within energy-based modeling frameworks.

AGI and the Evolution of AI

The Role of EBMs and LLMs in AI Development

  • Discussion on the potential cognitive aspects that contribute to Artificial General Intelligence (AGI), suggesting multiple elements may be involved.
  • Emphasis on the compatibility of new models with existing language models (LLMs), indicating a collaborative evolution rather than a complete replacement.
  • Recognition that different models, including EBMs (Energy-Based Models), will excel in specific tasks while being inadequate in others, highlighting the diversity within AI capabilities.
  • The concept of collective intelligence is introduced, comparing human society's strengths and weaknesses to those of AI systems, suggesting a symbiotic relationship.
  • Exploration of hybrid models combining EBMs and LLMs, proposing an agentic layer that could lead to advanced self-training and alignment processes.

Defining AGI: A Moving Target

  • Question raised about when AGI should be defined—whether it’s based on agents providing solutions or controlling complex systems like energy grids.
  • The classic definition revolves around generalization capabilities; current models are questioned for their ability to provide reliable answers across diverse data inputs.
  • Reflection on how perceptions of AGI have evolved over time; past benchmarks may now seem inadequate as technology progresses.
  • Anticipation that educational institutions will adapt their definitions and teachings regarding AGI as advancements continue to redefine what constitutes intelligence in machines.

Human Interaction with Robots

  • Inquiry into whether robots equipped with sensors require language models for effective communication with humans, emphasizing interaction needs.
  • Importance placed on understanding user requirements when designing robots; engineers must consider available data types for optimal functionality.
  • Discussion about various input sources for robots, such as visual data and environmental sensors, which inform their behavior and decision-making processes.
  • Purpose-driven design is highlighted; the robot's function influences its interaction level—e.g., caregiving robots need robust communication abilities alongside situational awareness.
  • Challenges noted regarding computational demands for real-time decision-making in robots compared to human capabilities.

Reflections on Human Abilities vs. AI Aspirations

  • Humorously suggests that humans' innate abilities might already represent a form of AGI we seek to replicate through technology.
  • Speculation about future generations questioning current reliance on human drivers illustrates evolving perspectives on technology's role in daily life.

How to Train Large Language Models (LLMs) Efficiently?

Understanding Model Training and GPU Utilization

  • The speaker discusses the current state of training LLMs, noting that they are still using a single GPU for initial experiments. They express uncertainty about how scaling will work in practice.
  • A key question raised is whether the architecture of LLMs is scalable and what parameters influence this scalability. The speaker emphasizes breaking down complexity without needing excessive GPUs.
  • The concept of phase transitions in LLM performance is introduced, where a critical mass of GPUs leads to significant changes in model behavior, enhancing its capabilities.
  • In hybrid models combining Energy-Based Models (EBMs) with LLMs, different regimes can dominate performance. If the EBM regime prevails, fewer GPUs may be needed compared to when the LLM dominates.
  • The use case significantly influences GPU requirements; real-time language processing likely necessitates more computational power from LLMs.

Phase Transitions and Their Implications

  • The speaker relates phase transitions to concepts in physics, highlighting their mathematical beauty and relevance to understanding model behavior during training.
  • An intriguing question arises about whether human cognition operates similarly to energy-based models. While there are theories regarding brain function, no definitive answers exist yet.
  • Good scientific theories must map well onto real data and make accurate predictions. Current models for visual cortex functioning have influenced AI development but do not fully explain brain operations.

Sustainability in AI Development

  • The discussion shifts towards sustainability in AI, particularly regarding energy efficiency and cost-effectiveness in running models.
  • Sustainable AI is defined as being both environmentally friendly and financially viable. This raises questions about optimizing resource usage within agentic systems involving EBMs or LLM hybrids.
  • There’s potential for optimizing agentic interactions between models to conserve resources like time and money while maintaining functionality.

Future Directions for AI Technology

  • The conversation touches on innovative applications of technology like Molten Book, which involves agents interacting creatively while managing compute costs effectively.
  • There's an emphasis on developing tailored solutions for businesses rather than general public access due to high operational costs associated with large-scale AI deployments.

This structured overview captures essential insights from the transcript while providing timestamps for easy reference back to specific discussions within the video content.

Understanding the Evolution of AI and Business Models

Resource Management in AI Development

  • The development of AI models takes a few days, allowing for precise resource management and control over security data.
  • Smaller models contribute to lower costs, contrasting with larger tech companies that invest heavily in generating media without clear purpose.

Historical Context and Future Prospects

  • The initial approach to business models involved giving users free access, but lessons learned from past experiences are now shaping better strategies.
  • Current research in AI is significantly more advanced than a decade ago, raising questions about future developments over the next 10 to 20 years.

Co-evolution of Humans and AI

  • There is an anticipated moment where human evolution will align with AI advancements, leading to new ways of thinking and working.
  • Concerns arise regarding whether reliance on AI tools will negatively impact learning processes among students.

Educational Implications

  • The speaker expresses curiosity about how educational systems will adapt to the integration of AI technologies.

Collaboration Opportunities with Kona Model

  • Discussion on how individuals can collaborate with Kona or create their own models using available services like ALF.
  • ALF serves as an agentic layer for code generation, initially developed as an internal tool before being tested publicly.

Safety and Transparency in Model Development

  • Emphasis on understanding model boundaries before public release to ensure safety and prevent misuse.

Advancements in Reasoning Energy Models

  • Inquiry into what has enabled reasoning energy models to advance now compared to previous decades; highlights the importance of latency space for effective task management.

Innovative Training Processes

  • New training methods include self-alignment capabilities for models, which were previously absent. This innovation stems from unexpected areas explored during PhD studies.

Company Launch Insights

  • The company was launched recently, receiving significant interest through emails and messages indicating strong engagement from potential users.

Exploring Curiosity and Education in AI

Collaboration with Experts

  • The team is partnering with professors and experts to create educational materials for their website, indicating a commitment to knowledge dissemination.
  • A small science team is focused on writing foundational papers about the mathematics behind Large Language Models (LLMs), showcasing an academic approach to understanding AI.
  • The discussion highlights the emergence of a new field in AI, specifically around EBM reasoning, emphasizing innovation and interdisciplinary collaboration.

Creativity and Knowledge Integration

  • The conversation reflects on how diverse knowledge can lead to innovative ideas, suggesting that creativity often arises from seemingly unrelated fields coming together.
  • There’s excitement about the potential growth of this new field, with participants expressing eagerness to engage further in discussions about it.

Communication and Teaching

  • The speaker shares insights into their role as a CEO, noting the repetitive nature of communication but finding ways to keep it engaging for audiences.
  • An appreciation for clear explanations is expressed, highlighting the importance of effective communication in complex topics like energy-based models (EBMs).

Future Implications of Energy-Based Models

  • EBMs are described as potentially transformative for AI thinking, indicating their significance in future developments within the field.
Video description

Most AI today is built to predict the next word. But intelligence doesn’t work that way. In this episode of The Neuron, hosts Corey Noles and Grant Harvey sit down with Eve Bodnia, Founder and CEO of Logical Intelligence, to explore energy-based models—a fundamentally different way to build AI systems. Unlike large language models, energy-based models don’t rely on tokens or next-word prediction. Instead, they reason over an energy landscape, allowing them to evaluate many possible solutions at once. Eve explains why this matters for spatial reasoning, planning, robotics, and safety-critical systems—and why language may be better treated as an interface, not the core of intelligence. We cover: • How energy-based models work (without the math) • Why LLMs hallucinate—and where constraints matter • How EBMs and LLMs can work together • What this approach means for the future of AI systems Learn more about Logical Intelligence: https://logicalintelligence.com/ Subscribe to The Neuron newsletter: https://www.theneuron.ai/ ➤ CHAPTERS 00:39 - Intro 01:17 - Logical Intelligence News: Yann LeCun Joins + Why Kona Matters 02:36 - Meet Eve Bodnia 08:01 - Why LLMs Hallucinate 13:00 - Kona’s Core Idea: Token-Free “Energy Landscape” Reasoning 14:52 - What AGI Means: Panning, Prediction & Verification 18:58 - How Energy Based Models Stay Correct 21:16 - Sudoku as a Benchmark for Non-Language Reasoning https://sudoku.logicalintelligence.com/ 22:15 - Why is Fast: Small Models, No Guessing 27:44 - Building Kona: Hybrid Architecture & Scaling 31:10 - Real-World Use Cases: Robotics & Scaling 40:00 - Scaling Theory: Phase Transition & GPU 47:29 - Business Model & Safety 51:14 - Why Energy Based Models Work Now 55:23 - Outro Hosted by: Corey Noles and Grant Harvey Guest: Eve Bodnia, Founder and CEO of Logical Intelligence Published by: Manique Santos Edited by: Kush Felisilda