| Episode | Status |
|---|---|
The robotics industry is on the cusp of its own “GPT” moment, catalyzed by transformative research advances. Enter Memo, the first general-intelligence personal robot, focused on taking on your chores...
Tony Zhao and Cheng Chi, co-founders of Sunday Robotics, discuss their journey from groundbreaking academic research to building Memo, the first general-purpose home robot. They explain how recent AI breakthroughs in diffusion policy, imitation learning, and scalable data collection have created robotics' 'GPT moment,' enabling robots that can generalize across environments. With over 10 million trajectories collected via their innovative glove system and 500+ data collectors, Sunday is preparing for a 2026 in-home beta program, targeting commercial availability by 2027-2028 at under $10k per unit.
The founders explain why robotics is experiencing its 'GPT moment' - the field has found scalable algorithms (diffusion policy, transformers) but hasn't yet scaled them to consumer products. Classical robotics required custom engineering for each task with no knowledge transfer, but new AI approaches enable generalization across tasks and environments.
Cheng explains how diffusion policy solved a critical problem in imitation learning - the ability to capture multiple ways of doing the same task while maintaining training stability. This breakthrough enabled scalable data collection from multiple untrained people, rather than requiring the original researcher to collect all data.
Tony describes how ALOHA transformed data collection from unintuitive VR headset teleoperation to a simple, reproducible setup that feels like playing a video game. Combined with ACT (Action Chunking Transformers), this enabled collecting truly dexterous manipulation data and proved transformers work for robotics when you have quality datasets.
Cheng developed UMI to solve the fundamental bottleneck: teleoperation restricted data collection to labs. By realizing you only need paired observation-action data (video + hand/gripper movement), he created a 3D-printed glove system with GoPros that enabled collecting 1,500 trajectories in two weeks across real-world environments, producing the first end-to-end model that generalizes to unseen environments.
After seeing the UMI robot work across Stanford campus, Tony and Cheng decided to start Sunday Robotics. The company has grown from two founders in an apartment to 30-40 people, bringing together expertise in mechanical engineering, supply chain, software, and controls to build a real consumer product, not just demos. Their mission: put a home robot in every home to eliminate mundane chores and give people time back for family and passions.
Sunday's design philosophy centers on 'what should a robot look like if it's ubiquitous?' They chose a friendly, cartoon-like face over a Terminator aesthetic. Every design decision prioritizes simplification to accelerate usefulness - like their 3-finger hand that combines three human fingers into one, reducing cost 3x while maintaining most functionality. AI's ability to correct for mechanical imprecision enables using compliant, low-cost actuators instead of expensive, precise industrial components.
Sunday plans a 2026 beta program putting various robot prototypes in real homes to learn how people interact with robots and what tasks they prioritize. Current prototypes cost $6k-$20k (mostly cladding at low scale), but will drop under $10k at a few thousand units through injection molding. Commercial availability targeted for 2027-2028, definitely not a decade away, with extremely high standards for safety, capability, and cost.
Sunday has collected nearly 10 million long-horizon trajectories (not just 'pick up a cup' but full tasks with walking and navigation) using 500+ people with gloves in the wild. The glove system went through 5 major versions with ~20 iterations each (~100 total iterations). Data quality became the critical factor at scale, requiring extensive engineering for automatic calibration, failure detection, and cleaning processes. Glove data surprisingly matches teleoperation quality while enabling more natural, dexterous movements.
Sunday sees RL working well for locomotion (easy to simulate rigid body dynamics) but believes imitation learning is faster for manipulation. The key insight: for manipulation, the behavior is easy to learn (close hand with right force) but the world is impossibly hard to simulate (transparent cup with liquid, hand deformation, lighting). For locomotion, it's flipped - world is easy to simulate but reactive behaviors are hard to specify. This makes imitation learning more sample-efficient for their use case.
Two major technical challenges remain: (1) figuring out the training recipe at scale now that sufficient data exists - Sunday is uniquely positioned with their data pipeline, and (2) hardware reliability as the learning team constantly pushes mechanical boundaries. The full-stack approach is essential because standards of 'good' are constantly evolving, making it impossible to work with external partners who can't keep up with changing requirements.
Sunday's demos showcase three capabilities: (1) Long-horizon mobile manipulation - cleaning entire messy table, loading dishwasher with precise force control to avoid breaking wine glasses, (2) Zero-shot generalization - tested in 6 Airbnbs with no additional data collection, grasping reflective silverware on transparent tables, and (3) Extreme dexterity - operating espresso machines and folding socks, tasks requiring fine force control impossible with numb teleoperation but natural with gloves where humans feel the forces.
Tony provides a framework for critically evaluating robotics demos: make zero assumptions. Ask if it's autonomous or teleoperated, whether it shows variation (different colored cups, different people), and count the number of interactions (more interactions = more failure opportunities = harder task). Humans naturally project capabilities onto demos, assuming a robot that hands one cup to one person can handle variations, but often it can only do that exact task.
Sunday Robotics: Scaling the Home Robot Revolution with Co-Founders Tony Zhao and Cheng Chi
Ask me anything about this podcast episode...
Try asking: