| Episode | Status |
|---|---|
Ilya & I discuss SSI’s strategy, the problems with pre-training, how to improve the generalization of AI models, and how to ensure AGI goes well.Watch on YouTube; read the transcript.Sponsors* Gemini ...
Ilya Sutskever discusses SSI's research-focused approach to superintelligence, arguing we've moved from the age of scaling to the age of research. He explores fundamental problems with current AI systems—particularly poor generalization compared to humans—and proposes that solving reliable generalization is key to safe superintelligence. The conversation covers technical approaches to alignment, the role of continual learning, and why SSI chose to focus on research over immediate product deployment.
Ilya examines why models perform exceptionally well on benchmarks but show limited economic impact. He proposes two explanations: RL training makes models too narrowly focused, and companies inadvertently train on eval-like tasks rather than real-world performance, leading to a generalization gap.
Using competitive programming as an example, Ilya explains how current AI training resembles a student who practices 10,000 hours on specific problems versus one who practices 100 hours but has better general understanding. This illustrates why models don't generalize as well as humans despite massive training.
Ilya discusses how evolution may have given humans advantages in vision and locomotion, but the superior learning ability in math and coding suggests a more fundamental difference. He explores the role of emotions as a built-in value function, citing cases of brain damage that eliminated emotions and destroyed decision-making ability.
Ilya argues that 2012-2020 was research, 2020-2025 was scaling, and now we're returning to research but with massive compute. The 'scaling' insight was powerful because one word told everyone what to do, but pre-training is running out of data and current approaches will plateau.
Ilya explains why SSI's $3B in funding is more competitive than it appears. Other companies spend heavily on inference, product engineering, and sales, while SSI focuses purely on research. He argues you don't need maximal scale to prove new ideas—transformers, ResNet, and AlexNet were all proven on modest compute.
Ilya discusses SSI's potential 'straight shot' strategy versus gradual release. He argues both approaches have merit: straight shot avoids market rat race and difficult trade-offs, while gradual deployment helps the world prepare and makes AI's capabilities concrete rather than abstract.
Ilya argues that the term 'AGI' exists as a reaction to 'narrow AI' (chess, checkers), and pre-training seemed to deliver it by improving everything uniformly. However, this overshoots the target—humans aren't AGIs but rely on continual learning. The future may be deploying eager learners, not finished systems.
Ilya envisions superintelligence as a learning algorithm that can master any job, not a pre-trained system that already knows everything. Deployment involves instances learning different roles, potentially leading to rapid economic growth. However, specialization through competition may prevent winner-take-all dynamics.
Ilya proposes that AI aligned to care about sentient life (not just humans) may be easier to build and more robust, since the AI itself will be sentient. He draws parallels to human mirror neurons and empathy. However, this raises questions about whether humans would remain in control if most sentient beings are AIs.
Ilya explores the puzzle of how evolution hard-coded sophisticated social desires (status, reputation) into the genome when these are high-level concepts requiring complex brain processing. Unlike simple desires (food smell), social motivations require the brain to piece together information, yet evolution encoded them quickly and reliably.
Ilya discusses how showing increasingly powerful AI will change behaviors—companies will collaborate on safety, governments will act, and AI companies will become more paranoid. He emphasizes the difficulty of imagining future AI power and the importance of making it concrete through demonstration rather than abstract discussion.
For long-term stability in a world with powerful AIs, Ilya reluctantly suggests human-AI integration (Neuralink-like technology) as a solution. Without it, humans become passive recipients of AI reports rather than participants. Integration allows wholesale transmission of understanding, keeping humans involved in AI decision-making.
Ilya Sutskever – We're moving from the age of scaling to the age of research
Ask me anything about this podcast episode...
Try asking: