| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
Adam Marblestone is CEO of Convergent Research. He’s had a very interesting past life: he was a research scientist at Google Deepmind on their neuroscience team and has worked on everything from brain...
Adam Marblestone explores fundamental differences between how brains and AI systems learn, arguing that evolution has encoded sophisticated loss functions and reward signals that enable humans to learn from far less data than current LLMs. He proposes that the brain uses omnidirectional inference rather than next-token prediction, and that understanding the 'steering subsystem' (innate reward functions) versus the 'learning subsystem' (cortex) is key to both AI capabilities and alignment. The conversation covers practical neuroscience research including connectomics, the potential for billion-dollar investments in brain mapping, and how these insights could inform the next generation of AI architectures.
Marblestone argues that evolution has built extensive complexity into loss functions rather than architecture, encoding specific curricula for different brain regions. Unlike ML's simple loss functions (next token prediction), the brain may use many specialized cost functions activated at different developmental stages, essentially 'Python code' for learning curricula that evolution has optimized over millions of iterations.
Addresses Ilya Sutskever's question about how genomes encode high-level desires. Marblestone explains Steve Byrnes' theory: the brain has a 'steering subsystem' (innate responses) and 'learning subsystem' (cortex). Evolution wires learned features to innate rewards by having the cortex predict steering subsystem responses, allowing abstract concepts like 'embarrassing a famous scientist' to trigger innate shame responses despite evolution never encountering podcasts.
Explores how neural networks 'amortize' Bayesian inference - instead of sampling possible causes for observations, they directly map observations to likely causes. Discusses whether brains do true probabilistic sampling or amortized inference, and how test-time compute in LLMs relates to this distinction. Digital minds can copy amortized solutions, potentially changing what's worth building in versus computing at runtime.
Explains how evolution achieves sample-efficient learning with minimal genomic information (3GB genome, small fraction for brain). The key is that reward functions are compact 'Python code' that exploit the learning subsystem's generalization. Single-cell atlases reveal the steering subsystem has vastly more diverse, bespoke cell types than cortex, suggesting most genomic complexity goes into specifying innate behaviors and reward circuits, not the learning algorithm itself.
Current LLMs use primitive RL (upweighting entire successful trajectories) compared to decade-old techniques like Q-learning with value functions. The brain likely implements both: basal ganglia doing simple model-free RL with finite action spaces, and cortex building world models that include predictions of rewards. Some theories suggest 'RL as inference' where you sample plans conditional on high reward.
Biological brains face energy constraints (20W, 200Hz) but may have advantages in co-locating memory and compute, unstructured sparsity, and natural stochasticity for sampling. The inability to copy or directly access brain states is a major disadvantage. Future AI hardware may adopt brain-like features (low voltage, co-located memory) while keeping digital advantages (copyability, random access).
Most cellular machinery in neurons is implementation detail for executing learning rules (like weight normalization or memory consolidation) rather than fundamentally new algorithms. However, some examples like cerebellar timing circuits suggest cells can store computational primitives (time constants) that would require complex wiring in pure connectionist models.
E11 Bio is developing optical connectomics to reduce mouse brain mapping costs from billions to tens of millions of dollars. Unlike electron microscopy, optical approaches enable 'molecularly annotated connectomes' showing synapse types and cell properties. The strategy mirrors the Human Genome Project's shift from expensive techniques to massively parallel sequencing, aiming for million-fold cost reductions through technology development before large-scale data collection.
Proposes using brain activity patterns as auxiliary training signals beyond simple labels. Instead of just 'cat/dog' labels, predict the full neural activity pattern humans exhibit when seeing cats. This could sculpt networks to represent information more like human brains, potentially improving generalization. Discusses practical challenges of collecting brain data at scale and whether this approach could create commercial value.
Marblestone's research assumes AI timelines longer than 5 years, as comprehensive neuroscience won't impact 2027 AGI scenarios. Estimates low billions needed for comprehensive brain mapping across species. Discusses funding mix: current work is philanthropy-based, but NSF tech labs, billionaire-backed moonshots, and potentially AI lab investment could accelerate progress. Key risk: moonshot companies might pursue flashy goals without doing fundamental science.
Adam Marblestone – AI is missing something fundamental about the brain
Ask me anything about this podcast episode...
Try asking: