Andrej Karpathy, a prominent voice in the AI community, recently delivered a stark assessment of large language models (LLMs), asserting, "LLMs don't work yet." This provocative stance, amplified through a summary by PJ on Twitter, has sparked considerable discussion among AI professionals, founders, and venture capitalists, probing the fundamental capabilities and limitations of current generative AI. The interview, dissected by the video's commentator, offers a crucial counterpoint to the prevailing hype, though the presenter ultimately remains optimistic about the trajectory of AI development.
Karpathy’s critique centers on several perceived cognitive deficiencies. He contends that LLMs lack sufficient intelligence, are not multimodal enough, cannot effectively use computers, and struggle with memory, failing to retain information previously provided. Such cognitive shortcomings, he suggested, could take "about a decade to work through." This perspective highlights a core tension in AI development: whether intelligence must be intrinsically embedded within the model's architecture or can be augmented through external systems.
