Fill out this form and an expert will reach out to schedule time to talk.
METR's study found AI tooling slowed developers down. We found something more consequential: Developers are completing a lot more tasks with AI, but organizations aren't delivering any faster.
The AI productivity debate took an unexpected turn in July when METR published findings that AI coding assistants made experienced developers 19% slower on complex tasks. Their controlled study of 16 seasoned open-source contributors sparked intense discussion across the developer community—and for good reason. Their findings challenge the widespread assumption that AI automatically boosts productivity.
METR's research deserves credit for bringing scientific rigor to a field often dominated by anecdotal claims. Their controlled methodology revealed important truths about AI's limitations with complex, brownfield codebases that require deep system knowledge and organizational context. Our telemetry from 10,000+ developers confirms this pattern: We see AI adoption consistently skewing toward newer hires who use these tools to navigate unfamiliar code, while more experienced engineers remain skeptical.
But for business leaders making AI investment decisions, METR's study answers only part of the question. While understanding individual task performance (and perception of AI) is valuable, the critical question for organizations isn't whether AI helps developers complete isolated assignments faster. It's whether AI helps businesses ship better software to customers more effectively.
{{ai-paradox}}
METR's controlled experiment studied 16 experienced developers from large open-source repositories, primarily using Cursor Pro with Claude 3.5 Sonnet, working on carefully designed tasks in an isolated environment. This approach yields clean, comparable data, but it falls short of capturing how software development actually happens in organizations.
Enterprise software delivery involves far more than individual coding speed. Code must be reviewed by teammates, pass through testing pipelines, navigate deployment processes, and integrate with work from dozens of other developers. A developer might very well complete some simple tasks faster with AI, but if that creates bottlenecks downstream, the organization sees no benefit.
Our analysis took a fundamentally different approach. Instead of controlled tasks, we analyzed telemetry from 1,255 teams and over 10,000 developers across multiple companies, tracking how AI adoption affects real work in natural settings over time. Rather than measuring isolated task completion, we examined the full software delivery pipeline, from initial coding through deployment to production. Our goal was to determine whether widespread AI adoption is correlated with significant changes to common velocity, speed, quality, and efficiency developer productivity metrics.
The results of Faros AI's study revealed a benefit METR's methodology couldn't capture: AI is enabling developers to handle more concurrent workstreams effectively and deliver significantly higher throughput.
Our data shows that developers on high-AI-adoption teams interact with 9% more tasks and 47% more pull requests per day. This isn't traditional multitasking, which research has long shown to be counterproductive. Instead, it reflects a fundamental shift in how work gets done when AI agents can contribute to the workload.
With AI coding assistants, an engineer can initiate work on one feature while their AI agent simultaneously handles another. They can start a refactoring task, hand it off to AI for initial implementation, then review and iterate while AI tackles the next item in the backlog. The developer's role evolves from pure code production to orchestration and oversight across multiple parallel streams.
This parallelization explains why we also found 21% higher task completion rates and 98% more merged pull requests, even as individual task speeds might not improve dramatically. For businesses, this distinction matters enormously. Organizations don't optimize for how quickly developers complete single tasks; rather, they optimize for how much valuable software they ship to customers.
Notably, while we identified this correlation with throughput and multi-tasking, the telemetry did not indicate a correlation between AI adoption and task or PR speed, as measured by their cycle times.
{{ai-paradox}}
Here's where our findings align with METR's concerns. Both studies reveal that AI introduces new complexities into software delivery:
These findings echo METR's observation that AI can create as many problems as it solves, particularly for complex work.
Our key insight: AI's impact depends entirely on organizational context. In METR's controlled environment, the organizational systems that could absorb AI's benefits simply didn't exist. In real companies, those systems determine whether AI adoption succeeds or fails.
Organizations can address these challenges through more strategic AI rollout and enablement, systematic workflow changes, and infrastructure improvements.
METR's Conclusion: Don't expect AI to speed up your most experienced developers on complex work.
Faros AI's Conclusion: Even when AI helps individual teams, organizational systems must change to capture business value.
Both approaches provide valuable data on where AI helps and where it doesn't. Any disconnect isn't surprising when you consider the fundamental differences in what each approach measures:
METR measured: Individual developer performance on isolated, well-defined tasks with no downstream dependencies.
Faros AI measured: End-to-end software delivery performance across interdependent teams with real business constraints.
METR's environment: 16 experienced developers, primarily Cursor Pro with Claude 3.5/3.7 Sonnet, controlled tasks, no organizational systems.
Faros AI’s environment: 10,000+ developers across all experience levels, multiple AI tools (GitHub Copilot, Cursor, Claude Code, Windsurf, etc.), natural work settings, full organizational context.
For engineering leaders, the Faros AI study demonstrates that AI is unleashing increased velocity but existing workflows and structures are blocking it. Developers don't work in isolation—they work within systems of code review, testing, deployment, and cross-team coordination. Whatever impact AI has on individual productivity only translates to business value if it successfully navigates these organizational processes.
{{ai-paradox}}
Our qualitative fieldwork and operational insights suggest that companies achieving meaningful AI gains are redesigning workflows to harness AI's unique strengths. This means:
Most importantly, successful organizations treat AI adoption as a catalyst for structural change. This approach focuses on how AI can reshape the organization of software development work, rather than on marginal gains for individual developers.
METR's research provides crucial insights into AI's limitations and the importance of human expertise in complex problem-solving and how AI tools will have to evolve to support brownfield codebases.
But the story doesn't end with individual task performance. The question for organizations is how to harness AI's strengths—particularly its ability to enable parallelization and handle routine work—while addressing its weaknesses through better systems, training, and workflow design.
The future of AI in software development won't be determined by whether it makes individual developers faster at isolated tasks. Organizations will be expected to adapt their systems, processes, and culture to leverage AI as a force multiplier for human expertise.
Both lab studies and real-world telemetry have roles to play in understanding that future. For engineering leaders making investment decisions today, the real-world evidence points to a clear conclusion: AI's business impact depends far more on organizational readiness and strategic AI deployment than previously understood.
The companies that recognize this distinction and invest accordingly will build the durable competitive advantages that matter in the age of AI-augmented software development.
Most organizations don't know why their AI gains are stalling. Faros AI can help. Book a meeting with an expert today.
How to build a high-impact program that drives real results.
What to measure and why it matters.
And the 5 critical practices that turn data into impact.