Want to learn more about Faros AI?

Fill out this form to speak to a product expert.

I'm interested in...
Loading calendar...
An illustration of a lighthouse in the sea

Thank you!

A Faros AI expert will reach out to schedule a time to talk.
P.S. If you don't see it within one business day, please check your spam folder.
Oops! Something went wrong while submitting the form.
Submitting...
An illustration of a lighthouse in the sea

Thank you!

A Faros AI expert will reach out to schedule a time to talk.
P.S. If you don't see it within one business day, please check your spam folder.
Oops! Something went wrong while submitting the form.

Translating AI-powered Developer Velocity into Business Outcomes that Matter

Discover the three systemic barriers that undermine AI coding assistant impact and learn how top-performing enterprises are overcoming them.

Neely Dunlap
Neely Dunlap
A dark blue background with the letters AI appearing to be walls, with a person walking through the letters with a flashlight, heading toward the A's dead end
7
min read
Browse Chapters
Share
August 6, 2025

Two weeks ago, we published the AI Productivity Paradox Report 2025, a landmark study that exposes the disconnect between the adoption of AI coding assistants and their organizational impact. Developer output increases, but engineering outcomes are flat. 

We also identified common AI adoption missteps that explain this paradox, including slow uptake, uneven usage, adoption that skews to less tenured engineers, and surface‑level tool usage. 

{{ai-paradox}}

Today, we examine another angle of the report: The systemic barriers that sap productivity momentum even after AI coding assistants reach critical mass, and what top‑performing companies are doing to beat the odds.

Why AI gains stall: Three systemic barriers

Developers using AI complete 98% more code changes and 21% more tasks. But these gains evaporate at the company level, where neither a positive nor a negative impact can be observed. 

Why is this happening? Three systemic barriers keep coming up in operational fieldwork: 

Summary infographic depicting the three barriers which appear to be stalling broader AI impact
Three barriers summary infographic

1. Downstream bottlenecks cancel out upstream gains

AI accelerates code creation, but review queues, brittle test suites, and sluggish release pipelines remain stuck in yesterday’s gear. By Amdahl’s Law, your delivery engine only moves as fast as its slowest stage—so faster coding simply piles more work onto the choke points.

2. Grassroots adoption lacks structure and scale

AI adoption is still driven by bottom-up experimentation, with developer enthusiasm undermined by a lack of centralized enablement. Developers spend time navigating tools without guidance, users receive little to no formal training, and there's rarely a strategy tailored to role or experience—resulting in inconsistent outcomes and uneven utilization. Without shared best practices and strong internal communities to socialize tips and recommendations, the organization struggles to convert adoption into lasting impact.

3. Directionless deployment drains ROI

Simply handing out licenses to Copilot, Claude Code, or Cursor isn’t a strategy. Without clear goals, usage policies, and change‑management plans aligned to business priorities, AI becomes “just another tool” instead of a catalyst for transformation.

What high-performing companies do differently

Some companies are seeing greater success and higher ROI from their AI investments. Their edge stems from  three mutually reinforcing practices:

a table explaining the three practices to achieve higher AI ROI
Three practices to achieve higher AI ROI

Blueprint for operationalizing AI engineering

As software teams transition from AI-assisted coding to agentic development, the complexity and autonomy of AI participation will increase. This creates new coordination demands, where code may be written, reviewed, or executed by agents working in parallel with humans.

Read the comprehensive research to discover practical steps that scale AI through the entire lifecycle, set the stage for agentic development, and ready your organization for the next phase of AI‑driven innovation.

Neely Dunlap

Neely Dunlap

Neely Dunlap is a content strategist at Faros AI who writes about AI and software engineering.

Connect
AI Is Everywhere. Impact Isn’t.
75% of engineers use AI tools—yet most organizations see no measurable performance gains.

Read the report to uncover what’s holding teams back—and how to fix it fast.
Discover the Engineering Productivity Handbook
How to build a high-impact program that drives real results.

What to measure and why it matters.

And the 5 critical practices that turn data into impact.
Want to learn more about Faros AI?

Fill out this form and an expert will reach out to schedule time to talk.

Loading calendar...
An illustration of a lighthouse in the sea

Thank you!

A Faros AI expert will reach out to schedule a time to talk.
P.S. If you don't see it within one business day, please check your spam folder.
Oops! Something went wrong while submitting the form.

More articles for you

Editor's Pick
AI
DevProd
10
MIN READ

Claude Code Token Limits: Guide for Engineering Leaders

You can now measure Claude Code token usage, costs by model, and output metrics like commits and PRs. Learn how engineering leaders connect these inputs to leading and lagging indicators like PR review time, lead time, and CFR to evaluate the true ROI of AI coding tool and model choices.
December 4, 2025
Editor's Pick
AI
Guides
15
MIN READ

Context Engineering for Developers: The Complete Guide

Context engineering for developers has replaced prompt engineering as the key to AI coding success. Learn the five core strategies—selection, compression, ordering, isolation, and format optimization—plus how to implement context engineering for AI agents in enterprise codebases today.
December 1, 2025
Editor's Pick
AI
10
MIN READ

DRY Principle in Programming: Preventing Duplication in AI-Generated Code

Understand the DRY principle in programming, why it matters for safe, reliable AI-assisted development, and how to prevent AI agents from generating duplicate or inconsistent code.
November 26, 2025