| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
Read the essay here.Timestamps00:00:00 What are we scaling?00:03:11 The value of human labor00:05:04 Economic diffusion lag is cope00:06:34 Goal-post shifting is justified00:08:23 RL scaling00:09:18 B...
A critical analysis of AI scaling assumptions and AGI timelines, arguing that current reinforcement learning approaches reveal we're further from AGI than short-timeline predictions suggest. The essay challenges the notion that pre-training scaling laws apply to RL, examines why AI hasn't achieved broad economic deployment despite impressive benchmarks, and proposes that continual learning—not recursive self-improvement—will drive the next phase of AI progress through distributed agent experience rather than singular breakthroughs.
Examines the logical tension between believing AGI is imminent while investing heavily in baking skills into models through RL. If models were truly human-like learners, the current approach of pre-training on every possible task (web browsers, Excel, robotics) would be unnecessary—humans don't need rehearsal for every tool they'll use. This contradiction is most visible in robotics, where true human-like learning would make the hardware problem trivial.
Uses a concrete example of a biologist identifying macrophages in slides to illustrate why current AI can't replace knowledge workers. While image classification is a 'solved' deep learning problem, the economic value isn't in building custom training pipelines for every lab-specific microtask. Human workers are valuable because they learn context-specific skills on the job without requiring bespoke training loops for each new situation.
Directly challenges the argument that slow AI adoption is due to normal technology diffusion lag. Argues that if models had human-level capabilities, they would diffuse much faster than human employees because they can instantly read entire company Slack/Drive, distill skills from other AI instances, and avoid the 'lemons market' problem of hiring. The fact that labs aren't earning trillions (knowledge workers' cumulative wages) proves capability gaps, not diffusion issues.
Defends the practice of updating AGI definitions as models improve but fail to deliver expected economic impact. While AI bulls correctly criticize bears for moving goalposts on reasoning/common sense, it's rational to revise definitions when models that would have seemed AGI-level in 2020 (like Gemini) still don't automate half of knowledge work. Predicts this pattern will continue: by 2030, models will have continual learning and earn hundreds of billions, but still won't have automated all knowledge work.
Challenges the assumption that reinforcement learning will scale as predictably as pre-training. While pre-training had clean, general trends across multiple orders of magnitude (though on a weak power law), there are no well-established public scaling laws for RL from verifiable reward. Cites Toby Borg's analysis suggesting a million-x scale-up in RL compute is needed for gains equivalent to a single GPT generation leap.
Proposes an alternative to recursive self-improvement singularity scenarios: a distributed intelligence explosion driven by continual learning agents deployed across jobs, bringing learnings back to a hive mind for batch distillation. Argues continual learning will progress like in-context learning did—gradual improvement over 5-10 years rather than sudden breakthrough. Predicts fierce competition between labs will prevent runaway advantages, as talent poaching, SF rumor mill, and reverse engineering have neutralized previous supposed flywheels.
An audio version of my blog post, Thoughts on AI progress (Dec 2025)
Ask me anything about this podcast episode...
Try asking: