The era of hand-engineered autonomous vehicle systems, once the industry standard, is rapidly giving way to a new paradigm of end-to-end deep learning. This profound shift, dubbed Autonomous Driving 2.0, represents a fundamental re-architecture of how intelligent machines perceive, plan, and navigate the physical world, promising scalability and generalization that eluded its predecessors.
Alex Kendall, CEO of Wayve, recently articulated this transformative vision in an interview with Pat Grady and Sonya Huang of Sequoia. Kendall highlighted the stark contrast between the traditional, modular robotics approach and Wayve's pioneering generalization-first strategy, emphasizing the pivotal role of foundation models and world models in accelerating autonomous capabilities.
In the nascent stages of autonomous vehicle development, the prevailing approach, or AV 1.0, was rooted in classical robotics. Companies meticulously hand-engineered distinct components for perception, planning, mapping, and control. This method, while seemingly logical, created massive C++ codebases, each module painstakingly crafted to address specific scenarios and environments. The inherent complexity and brittleness of this segmented architecture meant that deploying autonomous vehicles required extensive, often prohibitive, re-engineering for every new city or vehicle type, relying heavily on high-definition maps and expensive LiDAR systems.
