0
0
Summary & Insights

The bizarre, almost comical image of an AI that apologetically fixes a coding bug only to reintroduce the original mistake—oscillating forever between two flawed states—reveals a fundamental strangeness at the heart of today’s most advanced models. In his conversation, Ilya Sutskever suggests this isn’t just a quirky bug; it’s a symptom of a deep crisis in generalization. While AI systems ace narrow benchmarks, their ability to apply learning robustly in the messy real world lags far behind, creating a puzzling chasm between their apparent smarts and their practical, economic impact.

This gap points to what Sutskever believes is the central problem: we have entered a new “age of research.” The previous era, dominated by simply scaling up pre-training compute and data, is yielding diminishing returns. Now, with immense compute available, the bottleneck is no longer resources but ideas. The core challenge is that our models generalize dramatically worse than humans; they are astonishingly sample-inefficient and lack the reliable, common-sense robustness of a person learning a new skill. Sutskever uses the analogy of two students: one who crams for 10,000 hours on competitive programming problems becomes a champion at that narrow task, while another with innate understanding practices for only 100 hours. The second student, like a human, will have a far more adaptable and successful career. Our current reinforcement learning (RL) methods, he argues, often create the AI equivalent of the first student—overfitted and brittle.

The path forward, therefore, involves moving beyond the current “pre-training + RL” recipe to discover new learning principles. Key to this is developing sophisticated value functions—internal models that allow an AI to judge the quality of its actions mid-process, much like a human feeling they’re on the wrong track while solving a problem. This could dramatically improve learning efficiency. Furthermore, Sutskever reimagines the goal itself. Instead of a static “AGI” that knows everything, he envisions a future of continual learning agents—AI “apprentices” that can be deployed into the economy, learn any job on the fly, and merge their knowledge. This collective, learning intelligence could drive explosive economic growth and functionally become superintelligent without a single, monolithic “brain.”

Looking further ahead, Sutskever discusses the alignment and safety of such powerful systems. He predicts that as AI’s power becomes more palpable, companies and governments will become radically more cooperative and paranoid about safety. He floats the provocative idea that the most stable long-term outcome might involve humans merging cognitively with AI to stay meaningfully in the loop. For his own company, SSI, the strategy is to focus intensely on the research for reliable generalization, insulated from short-term market pressures, aiming to be a decisive voice when the transition to superintelligence truly begins.

Surprising Insights

  • The real “reward hackers” may be the human researchers themselves. Sutskever suggests the disconnect between benchmark performance and real-world utility arises because researchers instinctively design RL training to maximize performance on specific evaluations, inadvertently creating narrow experts rather than robust generalists.
  • AGI might be the wrong framing. The historical concept of “Artificial General Intelligence” as a system that can do any human job out-of-the-box is contrasted with a more plausible future: a super-efficient continual learner that starts with a foundational understanding but must learn each specific job, much like a human does.
  • Emotions as a value function. Sutskever draws a fascinating parallel between human emotions and a machine learning value function, proposing that our hardwired emotional responses serve as a robust, evolution-coded guide for decision-making in a vast array of novel situations—a mechanism current AI lacks.
  • Competition could foster AI diversity. He speculates that while pre-training on the same internet data makes all current LLMs similar, creating environments where multiple AI agents compete (not just collaborate) could naturally incentivize them to develop diverse, specialized problem-solving approaches.

Practical Takeaways

  • Look beyond benchmark scores. When evaluating AI progress or building applications, don’t be dazzled by performance on standardized tests. Prioritize testing in open-ended, real-world scenarios where robustness and generalizable reasoning are required.
  • Invest in better feedback mechanisms. For those training models, explore advanced value functions and internal critique mechanisms to provide richer, earlier feedback during training, moving beyond simple end-of-task rewards.
  • Design for continual learning. When architecting AI systems, assume they will need to learn and adapt continually from new experiences after deployment, rather than being static, pre-trained artifacts.
  • Prioritize research on generalization. The most significant breakthroughs will likely come from understanding and improving generalization—why humans learn so much from so little—not from marginal increases in scale or data.

Ilya Sutskever is the co-founder of OpenAI, is one of the most cited computer scientist in history with over 165,000 citations, and to me, is one of the most brilliant and insightful minds ever in the field of deep learning. There are very few people in this world who I would rather talk to and brainstorm with about deep learning, intelligence, and life than Ilya, on and off the mic.

Support this podcast by signing up with these sponsors:
– Cash App – use code “LexPodcast” and download:
– Cash App (App Store): https://apple.co/2sPrUHe
– Cash App (Google Play): https://bit.ly/2MlvP5w

EPISODE LINKS:
Ilya’s Twitter: https://twitter.com/ilyasut
Ilya’s Website: https://www.cs.toronto.edu/~ilya/

This conversation is part of the Artificial Intelligence podcast. If you would like to get more information about this podcast go to https://lexfridman.com/ai or connect with @lexfridman on Twitter, LinkedIn, Facebook, Medium, or YouTube where you can watch the video versions of these conversations. If you enjoy the podcast, please rate it 5 stars on Apple Podcasts, follow on Spotify, or support it on Patreon.

Here’s the outline of the episode. On some podcast players you should be able to click the timestamp to jump to that time.

OUTLINE:
00:00 – Introduction
02:23 – AlexNet paper and the ImageNet moment
08:33 – Cost functions
13:39 – Recurrent neural networks
16:19 – Key ideas that led to success of deep learning
19:57 – What’s harder to solve: language or vision?
29:35 – We’re massively underestimating deep learning
36:04 – Deep double descent
41:20 – Backpropagation
42:42 – Can neural networks be made to reason?
50:35 – Long-term memory
56:37 – Language models
1:00:35 – GPT-2
1:07:14 – Active learning
1:08:52 – Staged release of AI systems
1:13:41 – How to build AGI?
1:25:00 – Question to AGI
1:32:07 – Meaning of life

Leave a Reply

Lex Fridman PodcastLex Fridman Podcast
Let's Evolve Together
Logo