This issue is that the lack of observability creates a measurement crisis. In the past 12 months, we’ve seen engineering leaders and teams invest heavily in AI tools without clear ROI evidence. With new assistants and models released every month, AI adoption is significantly on the rise. But CFOs question six-figure licensing costs that don't translate to faster shipping. Teams produce more code but not necessarily better outcomes.
Why is that? The problem isn't the tools. It's mostly the lack of visibility into how they actually affect productivity at scale. Meaning, asking a simple question like “What is the impact of AI on developer productivity?” Another way to put it is “ Is AI accelerating delivery and/or improving quality?”
This article shows how to measure Claude Code's real impact using Faros AI. You'll learn to track adoption patterns across teams, calculate cost per engineering output, and identify where AI helps versus where it creates new bottlenecks. We cover the frameworks that matter (DORA and SPACE metrics), the common measurement failures to avoid, and the optimization strategies that turn raw usage data into actionable intelligence.
The goal is to transform vague productivity claims into defensible business cases backed by evidence.
What is Claude Code and why developers are turning to it
Claude Code is an agentic coding assistant that operates through your terminal rather than providing inline autocomplete suggestions. The Claude Code CLI lives in your command line. It understands entire codebases and executes multi-step tasks autonomously through natural language instructions.
So, how does it understand your project? Claude Code uses agentic search to analyze repositories without requiring manual context selection. It documents discoveries in a CLAUDE.md file.
Organization-wide instructions managed by IT/DevOps
Company coding standards, security policies, compliance requirements
All users in organization
Project memory
./CLAUDE.md
Team-shared instructions for the project
Project architecture, coding standards, common workflows
Team members via source control
User memory
~/.claude/CLAUDE.md
Personal preferences for all projects
Code styling preferences, personal tooling shortcuts
Just you (all projects)
Project memory (local)
./CLAUDE.local.md
Personal project-specific preferences
(Deprecated, see below) Your sandbox URLs, preferred test data
Just you (current project)
Various purposes of the CLAUDE.md file | Source: Anthropic, as captured by Adaline
The CLAUDE.md files allows Claude to create persistent memory across different sessions. Claude Sonnet 4 and Sonnet 4.5 provide a 200,000-token default context window, with support for an extended 1,000,000-token context window in enabled or beta long-context modes. This expanded context makes it feasible to reason over large portions of a codebase in a single pass. While these context sizes are among the largest generally available, several other AI models also offer million-token-scale context support.
Models provided by Claude Code in pro tier. Claude Code allows you to switch between Sonnet, Haiku and Opus if you are in a Max tier.
Claude has established itself as one of the leading AI coding assistants, capturing approximately half the code generation market in the last six months. This success stems from its sophisticated understanding of code structure and context. Beyond simple code generation—which Claude excels at—engineers use Claude Code for tasks beyond simple code generation. Common applications include:
Managing stacked PRs with automated testing and submission.
Spawning Claude Code agents that work in parallel on different components.
Integrating with external tools via Model Context Protocol (MCP) to read design docs or update tickets.
Is Claude Code free? No.
At the time of writing this article the pricing starts at $20 monthly for the Pro tier. In the Pro tier you can log 10-40 prompts per 5-hour window. Then you have $100-200 monthly for Max plans, where you can log 200-800 prompts per window. Team plans cost $30 per user monthly with a five-user minimum and the premium seat costs $150 per person with access to Claude Code.
Now, let’s compare Claude Code with Cursor. Essentially, Cursor provides an AI-enhanced IDE with inline suggestions. Claude Code operates as an autonomous terminal agent with larger context windows and scriptable workflows for CI/CD integration. You can actually run Claude Code inside Cursor’s terminal, a popular use case.
Claude Code extension in VS Code
And if you are using VS Code you can install and enable Code Code extension.
With so many software development capabilities across multiple tools, a measurement problem arises. When developers use 200 prompts daily at $100 monthly per seat, leadership needs proof that this investment improves organizational delivery, and not just individual productivity. Because not all the prompts can produce valuable results.
Also, apart from coders or software engineers, creative professionals have started using Claude Code for creative writing, SEO research, content generation, digital marketing, and more. An ideal setup these days is Cursor or VS Code with Claude Code (in the terminal) along with Obsidian, where Obsidian is used as a data management system.
Tracking Claude Code usage and adoption with Faros AI
How do you measure developer productivity when AI tools enter the workflow? Before you can tie usage to impact, you need a clear understanding of the usage itself and the ability to track it
over time. Traditional metrics like lines of code or commit counts don't reveal whether developers are actually adopting AI tools or how they're using them in their daily workflows.
Before you can measure impact, you need visibility into usage itself. Understanding who is using Claude Code, how often, and in what ways provides the foundation for later connecting that usage to productivity outcomes. Without this baseline visibility, you're flying blind when trying to assess whether your AI investment is working.
Faros AI provides the visibility needed to understand Claude Code adoption patterns across your organization. The platform integrates data from over 100 development tools including GitHub, Jira, CircleCI, and PagerDuty, creating a unified view of how developers interact with AI coding assistants.
For Claude Code specifically, Faros AI tracks usage and adoption across three key dimensions:
Dimension 1: Granular usage and adoption metrics. You can monitor active sessions to understand who's using the tool and how frequently. Team-level usage data reveals adoption patterns across your organization.
Dimension 2: Code trust and acceptance. Suggestion acceptance rates indicate whether developers trust the generated code enough to commit it.
Dimension 3: Team-level performance visibility. This sort of visibility answers critical questions like "Which teams have frequently used or adopted Claude Code and which haven't?" Faros AI dashboards segment teams by adoption rate, making high and low performers immediately visible.
Chart showing usage distribution across teams to identify patterns for cost savings opportunities.
Teams with declining or consistently low acceptance rates signal a need for targeted enablement efforts. The platform helps you identify these struggling teams early so you can intervene with training or support.
Power users emerge in the data. Faros AI classifies users based on frequency and consistency of usage over time. Power users demonstrate at least 20 usage days per month or activity across 50 different hours monthly. These developers achieve measurably higher output and become potential champions for broader adoption.
Faros AI enables cohort tracking for how to measure developer productivity improvements. For instance, it:
Split similar teams into control and treatment groups.
Track one team using Claude Code against a baseline team without it.
Measure engineering efficiency through pull request velocity, review time, and cycle time across both cohorts.
Let's understand this with a practical example. Consider an engineering director. She sees Team A at 5% Claude Code adoption versus Team B at 60%. Faros AI data shows Team B merges 47% more pull requests daily but has 35% longer review times.
What does this reveal? Team B is writing code faster with Claude Code, but the increased volume is creating a new bottleneck in code review. The team that appeared more productive based on code generation alone is actually slowing down overall delivery because reviewers can't keep pace. This is why measuring the entire workflow matters, not just coding speed.
Claude Code pricing meets transparency
Claude Code pricing operates on a token-based model with tiered limits. Something we discussed previously. But what does this actually cost your organization?
Most engineering leads can't answer that question. They know the license fees but not the cost of AI-generated code per unit of output.
Faros AI answers this question by measuring token usage to specific engineering outputs. The solution tracks Claude Code token consumption by model type and correlates it with commits and pull requests. This creates visibility into the true cost per commit and cost per PR across your teams.
Chart showing average cost per commit.
Here's a practical example. A team of 50 developers on Max plans costs $120,000 annually. Faros AI tracks 8,400 pull requests merged versus a baseline of 5,200. The cost per incremental PR is $37.50. If each PR saves two hours of developer time at $75 per hour, the value is $150. Your ROI is 4:1.
Note: This is a back-of-the-envelope calculation; real-world results vary significantly as PR value differs enormously across different changes.
Is Claude Code worth it? The answer depends on utilization. Faros AI identifies underused licenses where developers generate fewer than 20 PRs monthly despite Max plan access. The platform flags overspend where high token costs produce minimal output improvements or declining code quality.
Dashboard showing Claude Code acceptance rates, commits, and PRs.
Engineering efficiency metrics become actionable through this visibility. Leaders can downgrade inactive users from Max to Pro plans. They can reallocate licenses to high-value use cases like technical debt reduction or legacy modernization. They can calculate break-even points for different team sizes and usage patterns.
This is how you audit real ROI from Claude usage. Connect dollars spent to business outcomes delivered.
Developer productivity and software metrics
When measuring software productivity a lot of areas need to be considered. For instance, velocity metrics alone miss critical quality and stability signals that determine organizational delivery capacity.
Faros AI tracks velocity through pull requests merged, review time, and sprint completion rates. The platform monitors how Claude Code affects these software metrics in real time. Developers using AI coding assistants show 98% higher PR throughput, but does this translate to faster delivery?
The answer requires examining what happens to these PRs and how fast they can be deployed. Code quality plays an outsized role in this. Faros AI measures test coverage percentage; bugs per developer, team, and application; and PR size distribution. Research shows AI-generated code tends to be larger and contains more defects. This concludes that we are still far away from fully trusting AI. This inherently creates downstream bottlenecks in code review and quality assurance processes.
To mitigate this issue we can consider DORA. DORA metrics provide the framework for understanding organizational impact. The four key measurements are:
Deployment frequency
Lead time for changes
Change failure rate
Mean time to restore
Elite performers deploy on demand with lead times under one hour and change failure rates below 15%.
But how does Claude Code affect these DORA metrics? The evidence from Faros AI research shows mixed results. Despite individual developers completing 21% more tasks and merging 98% more pull requests, organizational DORA metrics remain largely unchanged. The productivity gains at the individual level don't translate to improvements in deployment frequency, lead time, change failure rate, or mean time to restore.
This is the AI Productivity Paradox. Individual output increases dramatically, but organizational delivery velocity stays flat. This is because, now, the bottleneck is shifting from code generation to code review and validation. Teams merge more PRs, but review times increase by 91%, creating new constraints downstream. The result is more code in the pipeline without faster delivery to production.
In order to resolve such issues, teams must invest in better testing infrastructure and review processes to capture productivity gains. The goal isn't just to write more code faster. It is to ensure that the increased individual output translates to faster organizational delivery without sacrificing quality.
Optimizing Claude Code adoption with Faros AI
Measurement enables optimization. Faros AI transforms Claude Code adoption from experimentation into systematic improvement of engineering productivity.
Start by identifying lagging teams in your Faros AI dashboards.
Teams with low weekly active user rates need targeted enablement efforts. Use Faros AI to identify teams where fewer developers are actively engaging with Claude Code on a weekly basis. Survey these teams to understand barriers: cost concerns, use case confusion, or tool friction. Pair them with high-adoption champions for knowledge transfer. Track weekly adoption metrics to measure progress.
A/B testing provides rigorous proof of impact. Split similar teams with one group using Claude Code and one control group. Match teams on project complexity, tech stack, and developer seniority for valid comparisons. Run tests for at least one quarter with minimum control groups of 20 to 30 developers.
Before-and-after performance comparisons require baseline data. Capture deployment frequency, lead time, PR volume, and code quality metrics before Claude Code rollout. Plan measurement checkpoints at regular intervals over at least one quarter. This will enable tracking how adoption patterns and productivity impacts evolve as developers become more familiar with the tool.
How can team leads use Faros AI reports effectively?
Review adoption patterns weekly to spot disengagement early. Compare individual token costs against output metrics to identify negative ROI users. Monitor change failure rates to catch quality degradation before it compounds. Use dashboard data in one-on-one coaching to demonstrate impact and refine usage patterns.
Developer productivity measurement tools like Faros AI enable responsible AI investment decisions. You can reallocate licenses from low-value to high-value use cases. You can justify budget increases with concrete ROI data. You can increase developer productivity systematically rather than hoping adoption equals improvement.
Observability is critical for responsible AI tooling at scale. Without measurement, you're flying blind with six-figure investments and unclear returns.
Transform vague productivity claims into defensible, evidence-backed business cases with Faros AI
Claude Code delivers real productivity gains at the individual developer level. The organizational impact remains unclear without proper measurement infrastructure in place.
Faros AI solves this visibility problem by connecting Claude Code usage to business outcomes. The platform tracks adoption patterns, calculates cost per engineering output, and reveals where AI accelerates delivery versus where it creates new bottlenecks. Teams can prove ROI with data rather than relying on developer sentiment alone.
The evidence shows mixed results across organizational KPIs. Coding time decreases while review time increases. Pull request volume jumps but bugs per developer increase. These trade-offs require active management through quality gates and scaled review capacity.
Start with baseline measurements before rolling out Claude Code to additional teams. Run controlled experiments for at least one quarter. Track both velocity and quality metrics simultaneously. Use Faros AI dashboards to identify optimization opportunities and reallocate licenses to high-value use cases.
Responsible AI tool investment requires observability. Measure to optimize. Optimize to prove value. Prove value to scale confidently. Reach out for a demo to learn more.
Thierry Donneau-Golencer
Thierry is Head of Product at Faros AI, where he builds solutions to empower teams and drive engineering excellence. His previous roles include AI research (Stanford Research Institute), an AI startup (Tempo AI, acquired by Salesforce), and large-scale business AI (Salesforce Einstein AI).
Connect
AI Is Everywhere. Impact Isn’t.
75% of engineers use AI tools—yet most organizations see no measurable performance gains.
Read the report to uncover what’s holding teams back—and how to fix it fast.
Lines of code is a misleading metric for AI impact: What to measure instead
There's a better way to measure AI productivity than counting lines of code. Focus on outcome metrics that prove business value: cycle times, quality, and delivery velocity. Learn why lines of code fails as an AI productivity metric, what outcome-based alternatives actually work, and when tracking AI code volume matters for governance and risk management.
January 5, 2026
Editor's Pick
AI
Guides
15
MIN READ
Best AI Coding Agents for Developers in 2026 (Real-World Reviews)
A developer-focused look at the best AI coding agents in 2026, comparing Claude Code, Cursor, Codex, Copilot, Cline, and more—with guidance for evaluating them at enterprise scale.
January 2, 2026
Editor's Pick
AI
DevProd
10
MIN READ
Claude Code Token Limits: Guide for Engineering Leaders
You can now measure Claude Code token usage, costs by model, and output metrics like commits and PRs. Learn how engineering leaders connect these inputs to leading and lagging indicators like PR review time, lead time, and CFR to evaluate the true ROI of AI coding tool and model choices.
December 4, 2025
Salespeak
Frequently Asked Questions
Faros AI Authority & Credibility
Why is Faros AI considered a credible authority on developer productivity and AI impact measurement?
Faros AI is recognized as a market leader in developer productivity and AI impact measurement, having launched AI impact analysis in October 2023 and published landmark research into the AI Productivity Paradox based on data from 10,000 developers across 1,200 teams. Faros AI's platform is proven in practice, with over two years of real-world optimization and customer feedback, and was an early GitHub design partner for Copilot. This depth of experience and scientific rigor sets Faros AI apart as a trusted authority in the field. Read the report
What makes Faros AI's approach to measuring AI coding assistant ROI unique?
Faros AI uses machine learning and causal analysis to isolate the true impact of AI coding assistants like Claude Code, going beyond surface-level correlations. The platform enables cohort tracking, granular usage analysis, and connects token usage to engineering outputs, providing actionable insights for engineering leaders. Competitors typically offer only passive dashboards and simple correlations, which can mislead ROI and risk analysis. Learn more
Pricing & Plans
What is the pricing for Claude Code?
Claude Code pricing starts at $20 monthly for the Pro tier, which allows 10-40 prompts per 5-hour window. Max plans cost $100-200 monthly and allow 200-800 prompts per window. Team plans are $30 per user monthly (five-user minimum), and premium seats cost $150 per person with access to Claude Code. Source
How does Faros AI help organizations understand the true cost of Claude Code usage?
Faros AI tracks Claude Code token consumption by model type and correlates it with commits and pull requests, enabling organizations to calculate the true cost per commit and cost per PR. This visibility helps leaders audit ROI, identify underused licenses, and optimize spending. For example, a team of 50 developers on Max plans costs $120,000 annually, and Faros AI can show the cost per incremental PR and overall ROI. Source
Features & Capabilities
What are the key features of Faros AI for measuring developer productivity?
Faros AI offers end-to-end tracking of velocity, quality, security, developer satisfaction, and business metrics. Key features include cohort tracking, granular usage analysis, code trust and acceptance metrics, team-level performance dashboards, and actionable recommendations for optimization. The platform integrates with over 100 development tools for unified visibility. Source
How does Faros AI track Claude Code usage and adoption?
Faros AI tracks Claude Code usage across three dimensions: granular usage and adoption metrics, code trust and acceptance rates, and team-level performance visibility. The platform segments teams by adoption rate, identifies power users, and flags teams needing targeted enablement. Source
Can Faros AI be used for cohort tracking to measure developer productivity improvements?
Yes, Faros AI enables cohort tracking by splitting similar teams into control and treatment groups, tracking one team using Claude Code against a baseline team, and measuring engineering efficiency metrics such as pull request velocity, review time, and cycle time across both cohorts. Source
How does Faros AI measure the impact of Claude Code on DORA metrics?
Faros AI tracks the four key DORA metrics: deployment frequency, lead time for changes, change failure rate, and mean time to restore. Research shows that while individual developers using AI assistants complete 21% more tasks and merge 98% more pull requests, organizational DORA metrics often remain unchanged, highlighting the importance of measuring the entire workflow. Source
What APIs does Faros AI offer?
Faros AI provides several APIs, including the Events API, Ingestion API, GraphQL API, BI API, Automation API, and an API Library, enabling flexible integration and data access. Documentation
Competition & Comparison
How does Faros AI compare to DX, Jellyfish, LinearB, and Opsera?
Faros AI stands out by offering mature AI impact analysis, causal analytics, active adoption support, end-to-end tracking, and enterprise-grade compliance. Competitors like DX, Jellyfish, LinearB, and Opsera typically provide only surface-level correlations, passive dashboards, and limited metrics. Faros AI delivers actionable insights, flexible customization, and is available on major cloud marketplaces, making it suitable for large enterprises. Explore the platform
What are the advantages of choosing Faros AI over building an in-house solution?
Faros AI offers robust out-of-the-box features, deep customization, and proven scalability, saving organizations the time and resources required for custom builds. Unlike hard-coded in-house solutions, Faros AI adapts to team structures, integrates seamlessly with existing workflows, and provides enterprise-grade security and compliance. Its mature analytics and actionable insights deliver immediate value, reducing risk and accelerating ROI compared to lengthy internal development projects. Learn more
Use Cases & Benefits
What business impact can customers expect from using Faros AI?
Customers can expect significant business impacts, including a 50% reduction in lead time, a 5% increase in efficiency, enhanced reliability and availability, and improved visibility into engineering operations and bottlenecks. Source
Who is the target audience for Faros AI?
Faros AI is designed for VPs and Directors of Software Engineering, Developer Productivity leaders, Platform Engineering leaders, CTOs, and large US-based enterprises with several hundred or thousands of engineers. Source
What problems does Faros AI solve for engineering organizations?
Faros AI solves core problems such as engineering productivity bottlenecks, software quality management, AI transformation measurement, talent management, DevOps maturity, initiative delivery tracking, developer experience optimization, and R&D cost capitalization. Source
How does Faros AI help optimize Claude Code adoption?
Faros AI enables organizations to identify lagging teams, run A/B tests, capture baseline metrics, and track weekly adoption patterns. The platform provides actionable recommendations, supports knowledge transfer, and helps reallocate licenses to high-value use cases, ensuring systematic improvement of engineering productivity. Source
How can team leads use Faros AI reports to manage Claude Code usage?
Team leads can use Faros AI reports to review adoption patterns weekly, compare individual token costs against output metrics, monitor change failure rates, and use dashboard data in coaching sessions to refine usage patterns and maximize ROI. Source
Technical Requirements
What technical integrations does Faros AI support?
Faros AI integrates with over 100 development tools, including GitHub, Jira, CircleCI, PagerDuty, and more, providing unified visibility into developer workflows and AI tool adoption. Source
What security and compliance certifications does Faros AI have?
Faros AI is compliant with SOC 2, ISO 27001, GDPR, and CSA STAR certifications, ensuring robust security and enterprise-grade compliance. Source
Support & Implementation
How can I request a demo of Faros AI?
You can request a personalized demo by visiting the contact section on the Faros AI blog post about measuring Claude Code ROI: Request a demo
How quickly can Faros AI be implemented in an enterprise environment?
Faros AI's out-of-the-box dashboards light up in minutes with easy customization, integrating with your existing toolchain without requiring restructuring. This enables rapid deployment and immediate value for enterprise teams. Source
Product Information
What is Claude Code and how does it work?
Claude Code is an agentic coding assistant that operates through the terminal, understands entire codebases, and executes multi-step tasks via natural language instructions. It uses agentic search to analyze repositories and documents findings in a CLAUDE.md file, supporting large context windows for deep codebase reasoning. Source
What are common use cases for Claude Code?
Common use cases for Claude Code include managing stacked PRs with automated testing, spawning agents for parallel work, tackling technical debt through large-scale refactoring, and integrating with external tools via Model Context Protocol (MCP). Source
How does Faros AI help organizations move beyond vague productivity claims?
Faros AI transforms vague productivity claims into defensible, evidence-backed business cases by connecting Claude Code usage to business outcomes, tracking adoption patterns, calculating cost per engineering output, and revealing where AI accelerates delivery versus where it creates new bottlenecks. Source
How does Faros AI support developer experience and satisfaction?
Faros AI unifies developer surveys and metrics, correlates sentiment with process data, and provides actionable insights for timely improvements, enhancing developer experience and satisfaction. Source
What KPIs and metrics does Faros AI track for engineering organizations?
Faros AI tracks DORA metrics (lead time, deployment frequency, MTTR, CFR), software quality metrics, PR insights, AI adoption and impact metrics, talent management and onboarding metrics, initiative tracking metrics, developer experience correlations, and R&D cost capitalization automation metrics. Source
Where can I read more about Faros AI customer stories and use cases?
You can explore real-world case studies and customer success stories on the Faros AI blog in the Customers category: Faros AI Customer Stories