Study Finds AI Coding Tools Reduce Productivity for Experienced Developers

A METR study reveals AI coding tools hinder experienced developers' productivity by 19%.

Key Points

  • • AI tools reduced experienced developers' productivity by 19%.
  • • Developers expected a 24% reduction in task completion time, but results were contrary.
  • • Less than 44% of AI-generated code was accepted without modification.
  • • Increased review time and reliance on AI outputs detracted from active coding time.

A recent study from Model Evaluation and Threat Research (METR) reveals that AI coding tools may hinder the productivity of experienced open-source developers, causing a notable 19% slowdown compared to their non-AI-using counterparts. The study involved 16 software engineers tasked with completing 246 coding tasks, half of whom were assisted by AI tools and half who were not.

Surprisingly, these seasoned developers had entered the study with expectations of enhanced efficiency, predicting a potential 24% reduction in task completion time when utilizing AI assistance. Instead, the results contradicted these expectations, showcasing that the AI-assisted group took longer to finish their tasks. The primary issue stemmed from the increased time spent reviewing AI-generated outputs, prompting the AI, and dealing with idle waiting periods, which outweighed any reductions in active coding time. Specifically, AI-assisted developers dedicated over 20% more time to reviewing outputs and spent about 9% of their total task time scrutinizing AI code, resulting in less than 44% acceptance of AI suggestions without modifications.

Nate Rush, a researcher involved in the study, noted the surprising nature of these findings, contemplating that the databases’ complexity might have played a role, as developers with less experience could potentially benefit more from AI tools. Critics have also suggested that advances in AI technology since the study's conclusion might alter results in future applications. Steve Newman, a co-founder of Google Docs, acknowledged the study's implications, stating that it critiques current AI limitations in complex coding scenarios even as the industry looks towards advancements to improve efficacy in developer workflows.

This study raises important questions about the current utility of AI coding tools, especially in highly structured coding environments where experience plays a crucial role. As developers assess the reliability of AI-generated code, it becomes increasingly clear that while AI has the potential to aid productivity, it may also introduce challenges that seasoned programmers must navigate carefully. The mixed results call for a more nuanced understanding of how AI tools can be best utilized by various skill levels in the evolving tech landscape.