Secret AI War Inside Apple

Secret AI War Inside Apple

They Might Be Self-aware: The Corporate War in AI

Introduction to the AI Landscape

  • Gary introduces the episode, highlighting a corporate rivalry between Claude and Gemini within Apple.
  • Discussion on how Google significantly undercut Anthropic's pricing by 90% to secure Apple's Siri contract, despite internal preferences for Claude among Apple engineers.

Insights on AI Development and Design

  • Hunter expresses frustration over Quad's design capabilities, suggesting that robots now possess aesthetic judgment previously thought unique to humans.
  • Commentary on the dominance of Google/Alphabet in the AI sector and concerns about OpenAI's long-term viability.

Vibe Coding vs. Traditional Coding Practices

  • Daniel defines "vibe coding" as using AI tools without deep understanding, contrasting it with traditional coding practices where expertise is essential.
  • Discussion on how large corporations are increasingly adopting agentic coding workflows, integrating tools like Claude or Codeex into their processes.

Evolving Definitions of Coding Practices

  • The conversation shifts towards "AI first" approaches where English descriptions lead to code generation through agents.
  • Daniel humorously identifies himself as a "vibe designer," acknowledging his limitations in traditional design while emphasizing his strengths in machine learning and NLP.

Current Tools and Preferences in AI Models

  • Acknowledgment that while there is still a role for manual coding, reliance on generative tools is becoming standard practice.
  • Hunter reveals his continued use of Anthropic’s services but also explores other models like Gemini for specific research needs.

Understanding BERT and Its Evolution

Introduction to BERT

  • BERT, named after the Sesame Street character, is a family of Transformer-based models designed for tasks like classification and named entity recognition.
  • While large language models are popular, they often struggle with traditional tasks such as classification and named entity recognition efficiently.

Modernization of BERT

  • The speaker discovered "modern BERT," an updated version of the original architecture featuring improvements like a longer context window.
  • This modernization was significant for the speaker, highlighting how tools can reveal previously unknown advancements in AI.

AI Model Preferences

  • The speaker discusses their non-monogamous approach to using various AI models from different companies based on specific needs.
  • Apple has partnered with Gemini for the next version of Siri; however, internal usage at Apple leans heavily towards Anthropic's Claude model instead.

Insights on Company Strategies

  • Despite promoting co-pilot widely, Microsoft also utilizes Claude internally due to its superior performance compared to alternatives.
  • Users often prefer Claude over co-pilot when given a choice, indicating its effectiveness in practical applications.

Financial Considerations in Partnerships

  • Apple's decision-making process involved evaluating partnerships with OpenAI and Google’s Gemini models due to financial implications.
  • OpenAI's competitive positioning led to a strained relationship with Apple, prompting them to explore other options like Anthropic and Google.

Cost Implications of AI Models

  • Apple required models that could run on their hardware while addressing security concerns; both Anthropic and Google were asked to adapt their models accordingly.
  • Although Anthropic showed promise, their high costs made them less favorable compared to Google's significantly cheaper offerings.

Future Predictions in AI Development

  • The speaker predicts that Google/Alphabet will emerge victorious in the ongoing "AI wars" due to their ability to subsidize costs effectively.
  • The disparity in pricing between Google's and Anthropic's models suggests that cost will play a crucial role in user adoption moving forward.

AI Model Comparisons and Cost Implications

Preferences for AI Models

  • The speaker discusses their preference for cloud models over Google Gemini Pro 3, citing better interaction quality with the former.
  • They highlight the high cost of premium subscriptions (e.g., $200/month), suggesting that many users may find it prohibitive.
  • Google is presented as a more affordable option, especially for casual users who don't require extensive usage.

Performance Metrics

  • Reference to Arena.ai's leaderboard shows Anthropic's Opus 4.6 leading with a score of 1.526, closely followed by Google's Gemini at 1.486.
  • Discussion on Apple's willingness to pay a premium for internal use while acknowledging they cannot extend this cost to all customers.

User Experience and Model Differences

  • The speaker emphasizes that different models feel distinct in user experience, suggesting varying "personalities" among them.
  • Personal preferences are shared: Claude is used for human-facing tasks while OpenAI’s CodeX 5.3 is preferred for coding tasks.

Accessing New Models

  • Mention of challenges in accessing benchmarks for OpenAI’s latest model due to restricted API access; only available through CLI tooling.
  • A humorous take on the timing of model releases, noting how OpenAI reacts to competitors' announcements rather than proactively announcing their own advancements.

Advanced Features and Cost Considerations

  • Introduction of agent teams in Opus 4.6 allows complex tasks to be broken down into subtasks handled by multiple agents simultaneously.
  • The speaker shares experiences using multiple agents effectively within one thread, which was previously unattempted due to complexity concerns.

Future Pricing Predictions

  • Concerns raised about future costs associated with using advanced AI models; current subsidized rates may not last long.
  • Speculation that monthly costs could rise significantly as reliance on these technologies increases, contrasting current prices around $200/month.

Understanding AI Subscription Models and Security Concerns

Subscription Costs and Value Perception

  • The Gemini Ultra plan is priced around $2,500 a month, which is deemed excessive for the average user. However, it may be justified for high-income professionals who rely on these tools for their jobs.
  • Companies are expected to subsidize these costs as they recognize the value of investing in advanced tools for their engineers, potentially increasing salaries by an additional $25,000 annually.

Performance and User Experience

  • There is a perception that performance can degrade over time with updates (e.g., Opus 4.5), leading users to question whether newer versions are less effective than previous iterations.
  • The company may intentionally reduce resources from older models to allocate them towards new releases, impacting user experience during transitions.

Token Management and Contextual Understanding

  • Opus 4.6 reportedly allows users to assign multiple tasks simultaneously without losing context, despite having a smaller context window compared to competitors.
  • A million token limit in Opus 4.6 enables extensive conversation history but also leads to increased costs when using API due to token-based billing.

Security Implications of AI Development

  • Criticism exists regarding "vibe coders" who may overlook essential security practices when creating software through AI-generated code.
  • There's a significant difference between quickly deploying minimum viable products versus developing secure applications that comply with industry standards like SOC 2.

Vulnerability Detection in Open Source Software

  • Opus 4.6 has improved its ability to identify vulnerabilities within existing open-source software, demonstrating this capability by finding over 500 issues in widely used programs.
  • Developers often rely on numerous dependencies from various libraries when building software; understanding this ecosystem is crucial for maintaining security across applications.

Understanding Dependency Management in Programming

The Joke of Dependencies

  • A humorous observation is made about the programming industry's reliance on a complex chain of dependencies, where even a simple function can lead to vulnerabilities if compromised.

Shifting Away from Dependencies

  • There is a growing trend among companies to eliminate unnecessary dependencies by utilizing agentic programming tools that isolate required functionalities.
  • An example is given regarding the need for reading Microsoft Word documents, emphasizing that only specific parts of libraries should be imported rather than entire packages with multiple dependencies.

Leveraging AI for Code Optimization

  • AI can assist in identifying and importing only the necessary code segments, a process referred to as "tree shaking," which removes extraneous code.
  • Developers can instruct AI to reimplement functions more efficiently while eliminating external dependencies, potentially improving performance and control over the system.

Concerns About Security Vulnerabilities

  • A warning is raised about the misuse of models like Claude Opus to discover vulnerabilities without disclosing them, posing risks similar to those faced by hackers.

The Cybersecurity Landscape

  • The discussion shifts towards cybersecurity threats such as ransomware and espionage, highlighting how malicious links can compromise systems.

The Evolution of AI Models

Introduction of Faster Models

  • A new model called 4.6 fast has been released, boasting double the speed of its predecessor while maintaining quality but at a higher cost.

Hardware Considerations

  • Curiosity arises regarding the hardware used for these faster models, particularly large chips known for their rapid inference capabilities despite being expensive.

Cost vs. Quality Trade-offs

  • The conversation touches on the trade-offs between cost, speed, and quality in accessing AI services; users must choose two out of three desirable attributes: cheap, fast, or good.

Task Execution Timeframes

Estimating Task Durations

  • An average wait time of 20 minutes for AI responses is noted; however, executing nine tasks took approximately 45 minutes due to comprehensive requirements including documentation and testing.

AI Model Usage and Trust in Automation

Initial Thoughts on Time Efficiency

  • The speaker reflects on the efficiency of using AI, noting that a task taking 20 seconds is preferable to spending months completing it manually. They express willingness to invest 45 minutes for significant time savings.

Experience with AI Models

  • For the first time, the speaker utilized an AI model (Claude) without reviewing the code, relying on a human-in-the-loop system for data verification. This marked a shift in their approach to automation.
  • The speaker describes adapting an existing tool through Claude to handle multiple tasks simultaneously, demonstrating trust in AI's output despite not checking the underlying code.

Confidence in AI Outputs

  • Previously cautious about testing outputs incrementally, the speaker now feels confident enough to execute full-scale tasks directly based on initial results from AI.
  • They mention feeling "very confident" about trusting the model's performance and even consider bypassing traditional checks entirely.

Comparison of AI Models

  • The speaker discusses Opus as a leading model for coding tasks but acknowledges that Codeex 5.3 may perform better. They note that while speed is essential, Opus remains their preferred choice due to its reliability.
  • Speed differences among models are highlighted; Claude is noted as faster than Codeex, which can significantly impact workflow efficiency.

Impact of Speed on Workflow

  • Instantaneous feedback from fast models changes how users think and work; having quick responses allows more time for critical thinking and multitasking during projects.
  • Despite advancements in speed, human involvement remains crucial in decision-making processes during software development to ensure quality and alignment with project goals.

Human Involvement vs. Automation

  • The necessity of human oversight is emphasized; rapid automation without adequate input could lead to subpar outcomes or misalignment with user needs.
  • Building software requires iterative processes involving reviews and adjustments rather than relying solely on automated outputs.

Aesthetic Improvements with New Models

  • The speaker notes improvements in design aesthetics when using Codeex 5.3 compared to previous versions, indicating enhanced capabilities in generating visually appealing outputs.
  • A specific example illustrates this point: after revisiting a design created by an earlier version (5.2), they found significant enhancements made possible by newer technology.

Emotional Response to Design Outputs

  • There’s an emotional reaction tied to design quality; outputs from different models evoke varying feelings regarding their authenticity and creativity.

This structured summary captures key insights from the transcript while providing timestamps for easy reference back to specific moments within the discussion.

Self-Awareness in AI: Are We Getting Closer?

The Nature of Self-Awareness in AI

  • Discussion on whether AI is becoming self-aware, with a suggestion that current models exhibit more self-awareness than before.
  • Introduction of an interview process for developing large language models, emphasizing the importance of clarifying questions to ensure alignment between ideas and implementation.

Effective Communication in AI Development

  • Importance of asking clarifying questions during the development phase to avoid misinterpretation of requirements and ensure the final product meets expectations.
  • Description of a project where multiple features were implemented through iterative questioning, highlighting the complexity involved in creating comprehensive systems.

Utilizing Feedback Loops

  • Acknowledgment that while initial implementations may not be perfect or enterprise-ready, they can still achieve significant progress through continuous feedback and adjustments.
  • Use of voice input tools (e.g., Super Whisper) to brainstorm ideas freely before structuring them into actionable plans.

Stress Testing Plans for Alignment

  • After generating a plan, the speaker emphasizes validating it by stress testing through specific questions to ensure all parties are aligned before moving forward.
  • Mention of using multiple-choice questions as a tool for assessing understanding and prioritization within the developed plan.

Engaging with Content and Future Discussions

  • Invitation for listeners to engage further with content across various platforms like YouTube and Spotify, hinting at future discussions about AI's trajectory towards self-awareness.
  • Encouragement for audience interaction through likes and comments, fostering community engagement around topics discussed.
Video description

*Claude vs Gemini: Apple secretly chose sides in the AI coding war.* Internally it’s Claude. Publicly it’s Gemini. The reason? Cost. Apple’s developers reportedly build with Claude 4.6. Siri leans toward Gemini. That split tells you everything about *Anthropic vs Google*, why *Claude is expensive*, why *Gemini is cheaper*, and how the AI coding war is really being decided. This week we break down: * What “vibe coding” actually means (and why enterprises are adopting agentic workflows) * Claude 4.6 and Claude agent teams changing software development * Why Apple may have ditched Claude for Gemini at scale * AI vulnerabilities, zero-day exploits, and the security arms race * OpenAI vs Claude vs Gemini — speed, cost, and model “taste” This isn’t just about chatbots. It’s about who builds the next generation of software — and who can afford to. If Google can subsidize and Anthropic can’t… If Claude builds better systems but Gemini scales cheaper… Who actually wins? They might be self-aware. But the companies deploying them definitely are. --- ⏱️ *CHAPTERS* 00:00 *Intro* 02:02 *Vibe Coding & Agentic AI Explained* – AI-first development, Claude Code & modern software workflows 08:48 *Claude vs Gemini: Apple’s AI Decision* – Siri partnership, internal Claude usage & Anthropic vs Google 17:50 *Claude 4.6 Agent Teams* – Parallel AI agents, coding speed & why Claude is expensive 23:24 *AI Vulnerabilities & Zero-Day Exploits* – Open source security flaws & AI-powered bug hunting 30:25 *OpenAI vs Claude vs Gemini* – Speed, cost, model “taste” & the AI coding war --- ⚡ *Listen now & get self-aware before your tools do.* 🎧 *Listen on Spotify:* https://open.spotify.com/show/3EcvzkWDRFwnmIXoh7S4Mb?si=3d0f8920382649cc 🍎 *Subscribe on Apple Podcasts:* https://podcasts.apple.com/us/podcast/they-might-be-self-aware/id1730993297 ▶️ *Subscribe on YouTube:* https://www.youtube.com/channel/UCy9DopLlG7IbOqV-WD25jcw?sub_confirmation=1 --- 📢 *Engage* Do you eat Claude or Gemini for breakfast? No MIXING. Drop your preference in the comments — and defend it. New here? Subscribe for twice-weekly AI chaos. 🧠 *They Might Be Self-Aware — but are we?* #AI #ClaudeVsGemini #Anthropic #Google #AICodingWar #TMBSA