AI Paradox: Coding Assistants Make Experienced Developers Slower
Study finds AI coding assistants actually slowed experienced developers by 19%, despite their belief of increased speed.
July 14, 2025

In a surprising contradiction to the prevailing narrative of artificial intelligence as a universal productivity booster, a recent study by the AI research nonprofit METR has found that experienced software developers were actually slower when using AI coding assistants. The research, which has sent ripples through the tech community, suggests a more complex relationship between AI tools and expert performance than previously understood, challenging the assumption that AI invariably enhances efficiency for all users in all scenarios. The study's findings are particularly striking because they contrast sharply with the developers' own perceptions of their productivity.
The METR study, titled "Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity," employed a randomized controlled trial to assess the impact of AI tools on a group of 16 seasoned open-source developers.[1][2][3] These developers, who had an average of five years of experience with the specific, large, and popular open-source projects they were working on, were tasked with completing 246 real-world issues, such as bug fixes and feature implementations.[4][5][3] Each task was randomly assigned to be completed either with or without the aid of AI tools, primarily Cursor Pro, a popular AI coding assistant, integrated with models like Claude 3.5/3.7 Sonnet.[4][2][3] The core finding was startling: on average, developers took 19% longer to complete tasks when they were allowed to use AI.[4][6][2] This slowdown occurred despite the developers' own beliefs; before the study, they predicted AI would speed them up by 24%, and even after completing the tasks, they estimated they had been 20% faster with AI assistance.[7][6][8]
Several factors have been identified as potential contributors to this unexpected slowdown. One major element appears to be the developers' deep familiarity and expertise with their respective codebases.[6][3] For these experts, the AI's suggestions may have been less helpful than they would be for a novice or someone unfamiliar with the project's architecture.[7][9] The study's authors noted that the AI's suggestions were often "directionally correct, but not exactly what's needed," forcing the experienced developers to spend valuable time reviewing, correcting, and refining the AI-generated code.[7][9] This process of verification and correction may have negated any time saved by the initial code generation. Furthermore, the complexity and sheer size of the open-source repositories, some with over a million lines of code, posed a significant challenge for the AI tools, which struggled to grasp the full context and implicit requirements of the projects.[6][5][3] The reliability of the AI's output was also a factor, with developers accepting less than 44% of the generated suggestions.[6][3]
The implications of the METR study are significant for the burgeoning AI industry, which has heavily invested in the narrative of AI as a revolutionary productivity tool for software development.[7][10] The findings challenge the broad claims of universal speedups and suggest that the benefits of AI are highly context-dependent.[7][11] While previous studies have shown significant productivity gains, they often used synthetic tasks or benchmarks that may not accurately represent the complexities of real-world software development in large, mature projects.[7][4] The METR study highlights a disconnect between performance on these benchmarks and practical application for seasoned professionals.[4] However, the authors caution against overgeneralizing their results, emphasizing that the observed slowdown might not apply to junior developers, those working on new or unfamiliar codebases, or in different software development settings.[9][11] Interestingly, despite the measured slowdown, many of the participating developers, and even the study's authors, continued to use the AI tools after the study, suggesting they may find value beyond pure speed, such as reducing cognitive load or making the development process more enjoyable.[9]
In conclusion, the METR study provides a critical and nuanced perspective on the impact of AI on software development. It demonstrates that for experienced developers working in familiar, complex environments, the current generation of AI coding assistants can paradoxically hinder productivity. The time spent verifying and correcting AI suggestions can outweigh the benefits of automated code generation. This research underscores the importance of rigorous, real-world testing of AI tools and suggests that the path to effectively integrating AI into expert workflows may be more intricate than a simple switch. It also points to a potential "skill ceiling," where learning to effectively collaborate with an AI partner is a skill in itself, one that takes time and experience to develop.[1][12] As the AI industry continues to evolve, understanding these contextual nuances will be crucial for developing tools that genuinely augment, rather than impede, human expertise.
Sources
[2]
[3]
[4]
[5]
[7]
[9]
[11]
[12]