| Episode | Status |
|---|---|
| Episode | Status |
|---|---|
If your MCP server has dozens of tools, it’s probably built wrong.You need tools that are specific and clear for each use case—but you also can’t have too many. This creates an almost impossible trade...
Alex Rattray, CEO of Stainless (the API company behind OpenAI and Anthropic's SDKs), reveals why current MCP implementations are fundamentally broken and proposes a radical solution: code execution sandboxes instead of traditional tool calls. He shares how his team uses AI internally with custom Git-based knowledge repositories, explains the critical context window and security challenges plaguing MCP adoption, and outlines a vision where AI agents write and execute code directly against APIs rather than making dozens of individual tool calls.
Alex explains APIs as the 'dendrites of the internet'—the fundamental connections that enable all modern software. He positions Stainless's core mission as making computer-to-computer communication easier, which naturally extends to enabling LLMs to interact with APIs through MCP (Model Context Protocol).
Alex reveals the fundamental problem with MCP: to replicate what humans can do in a dashboard, you'd need to expose hundreds of API endpoints as tools, which burns through the entire context window and confuses models. Current MCP servers are severely limited compared to their web UI counterparts, restricting AI capabilities to just a few operations.
Alex shares how he personally uses MCP servers (Notion, HubSpot, Gong, Postgres) to query business data across multiple systems. He maintains a Git repository where Claude stores curated notes, customer quotes, and SQL queries for future reference, creating a persistent knowledge base that reduces repeated MCP calls.
Alex outlines current best practices: keep tool counts low, make descriptions precise and specific, minimize input parameters, return minimal response data, and invest heavily in evaluation systems. He emphasizes the need for product management discipline to identify high-value use cases rather than trying to expose entire APIs.
Alex proposes a revolutionary approach: instead of dozens of MCP tools, give models just two—one to execute TypeScript code using the API's SDK, and one to search documentation. This reduces context usage to ~1,000 tokens upfront, eliminates pagination overhead, and leverages models' superior code-writing abilities over tool selection.
Alex argues that security must happen at the API layer through OAuth with granular permissions, not by limiting MCP tool exposure. The code execution sandbox should restrict network access to only approved API endpoints, preventing models from making unauthorized external connections.
Dan pushes Alex on go-to-market strategy, arguing that AI products that win (Stable Diffusion, Claude Code) are those willing to be YOLO early, while cautious approaches (DALL-E's private beta, Codex CLI's restrictions) fall behind. Individual developers need access now, even with security trade-offs.
Alex envisions a future where AI-written code for one-off tasks (like refunding a customer) becomes production software. When the same task repeats, the AI commits the code to the repo, turning exploratory chat interactions into permanent automation. Tool building becomes purely prompt engineering.
MCP Servers: Teaching AI to Use the Internet Like Humans
Ask me anything about this podcast episode...
Try asking: