Summary & Insights
The next decade of AI isn’t about understanding data that already exists, but about understanding new data—particularly the three-dimensional, physical world we inhabit. This core philosophy drives Fei-Fei Li and Justin Johnson, pioneers whose work from ImageNet to neural radiance fields (NERFs) has defined modern AI. They describe the current moment as a “Cambrian explosion,” where AI is expanding beyond language into pixels, video, audio, and, most fundamentally, spatial intelligence. For them, this evolution is a multi-decade continuum now reaching an inflection point, leading to their new venture, World Labs. The conversation traces the key unlocks—compute scaling, large datasets, and algorithmic breakthroughs—that brought us from supervised learning on labeled images to today’s generative models. However, they argue the next frontier is fundamentally different: moving from a one-dimensional, language-centric view of intelligence to machines that perceive, reason, and act within 3D space and time.
The discussion positions spatial intelligence as a capability as ancient and essential as language, critical for any entity—human, robot, or virtual agent—to interact with reality. While large language models (LLMs) operate on 1D sequences of tokens, spatial intelligence requires inherently 3D representations to model the physical world’s structure, physics, and affordances. This isn’t just about generating static 3D scenes; the vision encompasses dynamic, interactive worlds for gaming, education, and new media, as well as the operating system for augmented reality and robotics. The line between reconstructing the real world and generating imagined ones is blurring, thanks to techniques like NERFs, creating a unified approach to understanding and creating spatial environments.
World Labs aims to build the foundational models for this spatial layer of intelligence. The founders see the convergence of massively scaled compute, sophisticated algorithms, and new types of 3D-aware data making this possible now. The applications are vast, from drastically reducing the cost of creating rich, interactive 3D worlds to enabling seamless AR interfaces that could one day replace all traditional screens, to providing the “brain” for robots to navigate and manipulate the physical world. The team, including other luminaries like Ben Mildenhall (NERF) and Christoph Lassner (precursors to Gaussian splatting), embodies the multidisciplinary depth required for this deep tech challenge, combining expertise in machine learning, computer vision, graphics, and systems engineering.
Surprising Insights
- The staggering scale of compute growth: A training run for the landmark 2012 AlexNet model that took six days on two consumer GPUs would now take less than five minutes on a single state-of-the-art NVIDIA GB200 chip, highlighting the often-underestimated hardware explosion behind AI progress.
- Data’s role shifted from explicit labels to implicit structure: The key unlock for generative AI wasn’t just more compute, but learning to use data without explicit human labels. The era of supervised learning (e.g., ImageNet’s hand-labeled categories) gave way to systems that learn from the implicit structure in data, like the relationship between images and their alt-text on the internet.
- Reconstruction and generation are merging in computer vision: In the world of pixels, the historical fields of 3D reconstruction (deducing structure from photos) and generative AI are converging. Techniques like NERFs mean the same fundamental models can now both reconstruct a real scene and generate a novel one, a pivotal but under-discussed shift compared to the LLM narrative.
- Language is a “lossy” representation of the physical world: The hosts argue that language, being a purely human-generated, 1D signal, is an abstract and often lossy description of the rich, 3D world governed by physics. This makes spatial intelligence a fundamentally different problem from language modeling, requiring native 3D representations.
- Spatial intelligence could deprecate physical screens: In a future with advanced AR, spatial intelligence models that seamlessly blend digital information with the physical environment could make dedicated screens (phones, monitors, TVs) obsolete, as information is presented contextually directly in the user’s field of view.
Practical Takeaways
- When thinking about next-gen AI applications, consider the 3D and spatial context. The most transformative uses may not be in chat interfaces but in areas involving navigation, manipulation, design, and interaction with the physical or virtual 3D world.
- For technical builders, investing in understanding 3D data representations (like NERFs, Gaussian splats, and neural fields) is crucial. The underlying architecture of future models will likely prioritize these over sequences of tokens for spatial tasks.
- The convergence of AI and computer graphics is a major trend. Expertise in rendering, simulation, and graphics pipelines is becoming as valuable as traditional machine learning knowledge for building spatial intelligence systems.
- Start exploring how your domain could be changed by a shift from 2D to 3D media. Whether it’s education, entertainment, e-commerce, or professional training, the ability to generate or interact with interactive 3D worlds cheaply will open new frontiers.
- Leverage open models and tools in the 3D vision space to experiment. The field is moving rapidly, with many foundational techniques (like those underlying NERF) being accessible and implementable, providing a playground for prototyping spatial AI ideas.
Fei-Fei Li and Justin Johnson are pioneers in AI. While the world has only recently witnessed a surge in consumer AI, they have long been laying the groundwork for the innovations transforming industries today.
With the recent launch of Marble, the first product from their company World Labs, we are revisiting this conversation to explore the ideas that started it all. World Labs is focused on spatial intelligence, building Large World Models that can perceive, generate, and interact with the 3D world. Marble brings that vision to life, allowing anyone, from individual creators to major platforms, to generate 3D scenes directly from text or image prompts and turn complex 3D creation into a simple, creative process.
In this episode, a16z general partner Martin Casado talks with Fei-Fei and Justin about the journey from early AI winters to the rise of deep learning and multimodal AI. From foundational breakthroughs like ImageNet to the cutting-edge realm of spatial intelligence, they discuss the evolution of the field and what is next for innovation at World Labs.
Timecode:
0:00 – The Next Decade of AI
2:45 – Origins: Backgrounds of the Founders
6:50 – The Rise of Deep Learning & ImageNet
8:00 – Algorithmic Unlocks: Compute, Data, and Supervised Learning
12:00 – From Predictive to Generative AI
16:20 – The Journey to Spatial Intelligence
18:35 – Defining Spatial Intelligence
21:15 – 3D Data, Computer Vision, and Breakthroughs
23:15 – Reconstruction vs. Generation in Computer Vision
24:45 – Spatial Intelligence vs. Language Models
29:00 – Applications: Virtual, Augmented, and Physical Worlds
39:55 – Building World Labs: Team and Vision
41:55 – The North Star: Measuring Success in Spatial Intelligence
Resources:
Learn more about World Labs: https://www.worldlabs.ai
Learn more about Marble: https://Marble.WorldLabs.ai
Find Fei-Fei on Twitter: https://x.com/drfeifei
Find Justin on Twitter: https://x.com/jcjohnss
Find Martin on Twitter: https://x.com/martin_casado
Stay Updated:
If you enjoyed this episode, be sure to like, subscribe, and share with your friends!
Find a16z on X: https://x.com/a16z
Find a16z on LinkedIn: https://www.linkedin.com/company/a16z
Listen to the a16z Podcast on Spotify: https://open.spotify.com/show/5bC65RDvs3oxnLyqqvkUYX
Listen to the a16z Podcast on Apple Podcasts: https://podcasts.apple.com/us/podcast/a16z-podcast/id842818711
Follow our host: https://x.com/eriktorenberg
Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.
Stay Updated:
Find a16z on X
Find a16z on LinkedIn
Listen to the a16z Podcast on Spotify
Listen to the a16z Podcast on Apple Podcasts
Follow our host: https://twitter.com/eriktorenberg
Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.
Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

Leave a Reply
You must be logged in to post a comment.