The widely held belief that artificial intelligence universally enhances human intellect and output recently faced a stark challenge, as discussed by the hosts of Uncapped Pod. Far from making us inherently smarter or more efficient, a recent study suggests AI tools may, in specific contexts, lead to a surprising decline in productivity, particularly among experienced professionals.
One of the hosts of Uncapped Pod detailed findings from a METR uplift study, an evaluation focused on how AI is progressing. This particular research involved open-source developers, working on highly-starred repositories, who were subjected to a randomized controlled trial. The experiment aimed to measure the impact of AI assistance on their work.
In the trial, developers were issued random pull requests. Some worked on these tasks independently, while others utilized AI tools like Cursor and Claude 3.7. The researchers then measured both the developers’ self-perceived increase in speed and their actual productivity gains. The self-perception proved wildly optimistic; developers believed they were 20% more productive with AI.
The reality, however, was a striking contrast. “They were actually 19% less productive as a result of AI,” one host revealed, emphasizing the counterintuitive outcome. This productivity dip was not uniform but concentrated among the most skilled. “Especially the senior engineers… had the biggest decreases in productivity,” he noted, highlighting a critical insight: AI’s benefit might inversely correlate with an individual’s existing expertise in certain domains.
Several explanations for this peculiar phenomenon were proposed. A primary theory points to a “common failure mode in intellectual work where you default to procrastinating by doing a thing which seems productive but is like, is not moving the ball forward that much.” This suggests AI might enable a form of high-tech procrastination, where developers engage in seemingly busywork generated or facilitated by the AI, rather than tackling the core, complex problems that require deep human thought.
For instance, an engineer might spend excessive time refining AI-generated code that is ultimately suboptimal, or waiting for AI completions rather than independently problem-solving. This isn't necessarily a flaw in the AI itself, but rather a human behavioral adaptation to the presence of an always-available assistant. The ease of offloading certain cognitive burdens to AI might inadvertently disincentivize the rigorous, independent thought processes crucial for genuine innovation and efficient problem-solving, especially for those who already possess advanced skills. For now, the notion that AI unequivocally makes us smarter remains an open question.

