Jared Kaplan, Anthropic's Co-founder and Chief Science Officer, recently illuminated the foundational insights driving artificial intelligence's relentless march towards human-level capabilities. Speaking at AI Startup School in San Francisco, Kaplan, a former theoretical physicist, underscored how the predictable nature of AI scaling is not merely an engineering feat but an almost "physical truth" reshaping our understanding of intelligence itself. His discourse centered on the surprising regularity with which AI performance improves, a phenomenon he contends is as precise as any trend observed in physics or astronomy.
The remarkable progress in AI, Kaplan explained, stems from two core training phases: pre-training and reinforcement learning. Pre-training involves models learning to imitate human-written data and discern underlying correlations, while reinforcement learning optimizes these models based on human feedback, guiding them toward "helpful, honest, and harmless" behaviors. Crucially, Kaplan highlighted that both phases exhibit clear scaling laws, meaning that as compute power, dataset size, and model parameters increase, performance improves in a highly predictable manner.
