World's Top Researcher on AI, LLMs, and Robot Intelligence

Robotics is entering a new era where general-purpose foundation models—similar to how ChatGPT handles any language task—will enable any robot to perform any physical task in any environment. The key insight: rather than building specialized robots for specific jobs (like dishwashing), the path forwa

March 31, 2026 1h 15m
Invest Like The Best

Key Takeaway

Robotics is entering a new era where general-purpose foundation models—similar to how ChatGPT handles any language task—will enable any robot to perform any physical task in any environment. The key insight: rather than building specialized robots for specific jobs (like dishwashing), the path forward is training one intelligent system that understands physical interaction fundamentally. This approach leverages diverse data sources and prior knowledge to handle both routine tasks and unexpected situations with common sense, much like humans do.

Episode Overview

Sergey Levine, co-founder of Physical Intelligence, discusses the company's mission to develop foundation models that can control any robotic system to perform any task. The conversation explores why building general-purpose robotic intelligence is actually easier than creating narrow, task-specific robots—mirroring how language models evolved. Key topics include: the "scarecrow problem" (robots need brains, not just bodies), how vision-language-action models bring web-scale knowledge to physical tasks, the role of reinforcement learning in exceeding human performance, and the path to a "Cambrian explosion" of robotic applications once the intelligence platform exists.

Key Insights

General-Purpose Beats Specialized in Robotics

Just as language models became more effective by solving natural language in its full generality rather than targeting narrow tasks like translation, robotic intelligence should tackle physical interaction broadly. Training models across many tasks and robots builds fundamental understanding of physics, causality, and object interaction—making it faster to master new skills than starting from scratch for each application.

The Missing Piece: Common Sense from Language Models

Historically, robotic learning struggled with "long-tail" scenarios—unusual situations the robot never experienced. The breakthrough: multimodal language models contain vast world knowledge from web-scale training. By using chain-of-thought reasoning (the robot literally "thinks" about what it should do before acting), systems can apply semantic knowledge to novel physical situations, handling edge cases with common sense.

Data Efficiency Through Foundation Models

Unlike narrow systems requiring massive data collection for each new task, foundation models pre-trained on diverse robotic data need far less task-specific training. The model develops "physical understanding"—intuitive grasp of what will happen in unfamiliar situations—enabling rapid skill acquisition similar to how humans quickly master new physical tasks.

Moravec's Paradox and Machine Learning

Things easy for humans (picking up a cup) are hard for robots, while human-difficult tasks (calculus) are machine-easy. However, machine learning shifts this equation: domains where collecting data is straightforward become easier over time, even if physically intricate. The remaining challenges are tasks requiring multi-level reasoning and connecting physical skills to web knowledge.

Surprising Generalization Across Embodiments

The same model works across radically different robot bodies—multi-fingered hands, different degrees of freedom, various form factors—without being explicitly told what robot it's controlling. This suggests the core challenge is understanding physical interaction, not adapting to specific hardware configurations.

Notable Quotes

"Fundamentally, the goal of physical intelligence is to develop robotic foundation models that can control basically any embodied system to do any task."

— Sergey Levine

"We believe that doing it at the full level of generality might actually in the long run be easier than trying to special case very specific narrow application domains."

— Sergey Levine

"People can master new skills very very rapidly because we understand physical interaction—we can intuitively grasp what's going to happen in this new unfamiliar situation and let us bootstrap things really really quickly."

— Sergey Levine

"Effective robotic learning, effective generalization isn't actually the optimal way to have like a really exciting demo. The way to have a really exciting demo is to pick a really cool task, control everything else in the environment, and just make it work in that one setting."

— Sergey Levine

"I think something like that might happen in the world of robotics but it can't happen today because if you want to put together some cool new robotics application, you kind of have to build this monstrous stack and you need to basically solve the intelligence problem."

— Sergey Levine

Action Items

  • 1
    Prioritize Learning Over Immediate Results

    When approaching complex problems, focus on building fundamental understanding across diverse scenarios rather than optimizing for quick wins in narrow domains. Like foundation models, invest in broad knowledge that makes future challenges easier to tackle.

  • 2
    Leverage Existing Knowledge Reservoirs

    Before collecting massive task-specific data, identify and incorporate relevant knowledge from adjacent domains. Chain-of-thought approaches—explicitly reasoning through problems using prior knowledge—can handle novel situations more effectively than pure experience.

  • 3
    Design for Data Collection from Day One

    Build systems that can gather useful data while being deployed, even if initially imperfect. Like Tesla's fleet learning approach, create feedback loops where real-world usage continuously improves the model rather than waiting for complete training datasets.

  • 4
    Question Moravec's Paradox in Your Domain

    Identify what seems "obviously easy" but is actually hard, and vice versa. In any field, human cognitive biases about difficulty can mislead strategic decisions—systematically test assumptions about what will be challenging to automate or scale.

  1. Podcasts
  2. Browse
  3. World's Top Researcher on AI, LLMs, and Robot Intelligence