Do AI Coding Assistants Slow Down Experienced Developers? Surprising Study Results

A recent randomized controlled study by the non‑profit AI research group METR found that, contrary to the widely held belief that AI coding tools boost developer speed by about 20%, experienced open‑source developers actually took 19% longer to complete real‑world tasks when using such tools, revealing a gap between perceived and actual productivity gains.

AI Frontier Lectures
AI Frontier Lectures
AI Frontier Lectures
Do AI Coding Assistants Slow Down Experienced Developers? Surprising Study Results

Research Motivation

Benchmarking studies of AI‑assisted coding often use isolated, synthetic tasks that ignore real‑world context and human interaction. METR argued that in‑situ, empirical experiments with experienced developers are needed to assess how generative AI actually influences software‑development productivity.

Methodology

METR recruited 16 seasoned open‑source contributors (average five years of experience, each maintaining repositories with >22 k stars and millions of lines of code). The participants supplied 246 real‑world issues from their own projects, covering bug fixes, feature additions, and refactorings. Each issue was randomly assigned to one of two conditions:

AI‑allowed: developers could use any coding assistant (the majority used Cursor Pro with Claude 3.5/3.7 Sonnet).

AI‑blocked: developers worked without generative assistance.

Developers recorded their work sessions, reported total elapsed time for each task, and were compensated at $150 USD per hour.

Key Findings

When AI assistance was permitted, developers spent on average 19 % more time to complete tasks than when AI was prohibited, directly contradicting the common expectation of a ~20 % speed‑up. Despite the slowdown, participants still believed AI would make them 24 % faster and remained confident in a 20 % productivity gain.

Time‑budget analysis showed that AI users spent less time writing code and searching for information, but allocated more time to:

crafting prompts,

waiting for model outputs,

reviewing generated code, and

idle periods.

Factor Analysis

METR examined 20 experimental attributes (e.g., model version, prompt length, task difficulty, interruption frequency). Five factors were identified as likely contributors to the observed slowdown, while eight factors showed mixed or ambiguous effects.

Robustness Checks

The slowdown persisted across alternative outcome metrics (e.g., lines of code changed, number of commits), different estimation techniques, and subgroup analyses (e.g., by task type or developer experience).

Limitations and Future Outlook

The study’s scope is limited to a small, highly skilled sample and to specific AI tools (Cursor Pro + Claude 3.x). Results may not generalize to all software‑engineering contexts or to developers with lower AI experience. METR acknowledges that newer or more specialized models could yield different outcomes and calls for continued diversified evaluation methods to track AI’s evolving impact on productivity.

Reference

Paper title: Measuring the Impact of Early‑2025 AI on Experienced Open‑Source Developer Productivity

URL: https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf

Code example

收
藏
,
分
享
、
在
看
,
给
个
三
连
击呗!
Original Source

Signed-in readers can open the original source through BestHub's protected redirect.

Sign in to view source
Republication Notice

This article has been distilled and summarized from source material, then republished for learning and reference. If you believe it infringes your rights, please contactadmin@besthub.devand we will review it promptly.

AIAI Toolsopen-sourcedeveloper productivityEmpirical Study
AI Frontier Lectures
Written by

AI Frontier Lectures

Leading AI knowledge platform

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.