| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
Kathleen Fisher and Byron Cook dive into automated reasoning and formal verification as tools for building truly secure software systems. PSA for AI builders: Interested in alignment, governance, or A...
Kathleen Fisher and Byron Cook explore how formal methods and automated reasoning can secure critical software systems against AI-enabled cyber threats. They explain how these mathematical techniques provide provable security guarantees, discuss AWS's decade of applying formal verification to cloud infrastructure, and reveal how generative AI is accelerating both proof discovery and secure code generation. The conversation culminates in examining AWS's automated reasoning checks for AI agents and the potential for a 'great software rewrite' where AI-generated code achieves superhuman security levels.
Fisher and Cook assess how AI is amplifying cyber threats across all skill levels and attack stages, from script kiddies to nation-state actors. They explain that AI helps attackers at every point in the cyber kill chain while also noting the optimistic potential for AI to strengthen defenses through formalization and automated reasoning.
The guests provide a foundational explanation of formal methods as algorithmic proof search following rigorous logical rules. They clarify the spectrum from simple type checking to full functional correctness, emphasizing that all proofs rest on assumptions and that the goal is raising assurance rather than achieving absolute certainty.
Fisher recounts the landmark HACMS program where formal methods secured a military helicopter against red team attacks. The system used SEL4 hypervisor, parser generators, and architecture modeling to prove system-wide security properties, withstanding attacks even during flight with test pilots aboard.
Cook details AWS's systematic application of formal methods since 2014, including tools for customers to verify their configurations and internal proofs of critical infrastructure. He explains how these efforts are now connecting into a comprehensive security framework, with the policy interpreter being called over a billion times per second.
Cook illustrates the fundamental difficulty of formal methods: defining what you want to prove. Using the example of 'all data at rest is encrypted,' he shows how seemingly simple requirements require extensive iteration to precisely specify what encryption means, what 'at rest' means, and what edge cases exist.
The guests explain how LLMs are transforming formal methods by finding inductive invariants and ranking functions that previously required PhD-level human insight. They describe the hierarchy of proof complexity and how generative AI excels at the hardest parts while combinatorial solvers handle verification.
Fisher and Cook outline how AI-generated code can achieve superhuman security levels through a flywheel of proof-based training. They explain that verification is easier than generation, enabling AI to generate proofs, validate them, and use successful proofs as training data - with formal methods providing reward signals for secure coding.
Cook demonstrates how AWS's automated reasoning checks translate natural language policies into formal logic for verifying AI agent outputs. The system uses multiple translations with theorem proving to achieve 99% accuracy, employing active listening to resolve ambiguities and creating a human-in-the-loop formalization process.
Fisher argues that technology for a society-wide secure software rewrite exists now, but motivation is the limiting factor. She compares the situation to Y2K, noting society can mobilize when sufficiently motivated, while acknowledging we've been 'boiling frogs' for 20 years of increasing cyber threats without adequate response.
The conversation concludes with a novel concern: as AI agents become perfectly consistent at policy enforcement through formal methods, we may lose valuable human flexibility to make exceptions when official policy doesn't fit specific situations. This represents a new class of problem emerging from successful AI alignment.
The Great Security Update: AI ∧ Formal Methods with Kathleen Fisher of RAND & Byron Cook of AWS
Ask me anything about this podcast episode...
Try asking: