Fill out this form to speak to a product expert.
Track Claude Code usage, adoption, and engineering impact using Faros AI’s observability platform.

Engineers and developers using Claude Code report dramatic individual gains. Some cite a 164% increase in story completion. Others nearly doubled their pull request merge rates. Yet organizations struggle to see these improvements materialize in overall delivery metrics.
This issue is that the lack of observability creates a measurement crisis. In the past 12 months, we’ve seen engineering leaders and teams invest heavily in AI tools without clear ROI evidence. With new assistants and models released every month, AI adoption is significantly on the rise. But CFOs question six-figure licensing costs that don't translate to faster shipping. Teams produce more code but not necessarily better outcomes.
Why is that? The problem isn't the tools. It's mostly the lack of visibility into how they actually affect productivity at scale. Meaning, asking a simple question like “What is the impact of AI on developer productivity?” Another way to put it is “ Is AI accelerating delivery and/or improving quality?”
This article shows how to measure Claude Code's real impact using Faros AI. You'll learn to track adoption patterns across teams, calculate cost per engineering output, and identify where AI helps versus where it creates new bottlenecks. We cover the frameworks that matter (DORA and SPACE metrics), the common measurement failures to avoid, and the optimization strategies that turn raw usage data into actionable intelligence.
The goal is to transform vague productivity claims into defensible business cases backed by evidence.
Claude Code is an agentic coding assistant that operates through your terminal rather than providing inline autocomplete suggestions. The Claude Code CLI lives in your command line. It understands entire codebases and executes multi-step tasks autonomously through natural language instructions.
So, how does it understand your project? Claude Code uses agentic search to analyze repositories without requiring manual context selection. It documents discoveries in a CLAUDE.md file.
The CLAUDE.md files allows Claude to create persistent memory across different sessions. Claude Sonnet 4 and Sonnet 4.5 provide a 200,000-token default context window, with support for an extended 1,000,000-token context window in enabled or beta long-context modes. This expanded context makes it feasible to reason over large portions of a codebase in a single pass. While these context sizes are among the largest generally available, several other AI models also offer million-token-scale context support.

Claude has established itself as one of the leading AI coding assistants, capturing approximately half the code generation market in the last six months. This success stems from its sophisticated understanding of code structure and context. Beyond simple code generation—which Claude excels at—engineers use Claude Code for tasks beyond simple code generation. Common applications include:
Is Claude Code free? No.
At the time of writing this article the pricing starts at $20 monthly for the Pro tier. In the Pro tier you can log 10-40 prompts per 5-hour window. Then you have $100-200 monthly for Max plans, where you can log 200-800 prompts per window. Team plans cost $30 per user monthly with a five-user minimum and the premium seat costs $150 per person with access to Claude Code.
Now, let’s compare Claude Code with Cursor. Essentially, Cursor provides an AI-enhanced IDE with inline suggestions. Claude Code operates as an autonomous terminal agent with larger context windows and scriptable workflows for CI/CD integration. You can actually run Claude Code inside Cursor’s terminal, a popular use case.

And if you are using VS Code you can install and enable Code Code extension.
With so many software development capabilities across multiple tools, a measurement problem arises. When developers use 200 prompts daily at $100 monthly per seat, leadership needs proof that this investment improves organizational delivery, and not just individual productivity. Because not all the prompts can produce valuable results.
Also, apart from coders or software engineers, creative professionals have started using Claude Code for creative writing, SEO research, content generation, digital marketing, and more. An ideal setup these days is Cursor or VS Code with Claude Code (in the terminal) along with Obsidian, where Obsidian is used as a data management system.
How do you measure developer productivity when AI tools enter the workflow? Before you can tie usage to impact, you need a clear understanding of the usage itself and the ability to track it
over time. Traditional metrics like lines of code or commit counts don't reveal whether developers are actually adopting AI tools or how they're using them in their daily workflows.
Before you can measure impact, you need visibility into usage itself. Understanding who is using Claude Code, how often, and in what ways provides the foundation for later connecting that usage to productivity outcomes. Without this baseline visibility, you're flying blind when trying to assess whether your AI investment is working.
Faros AI provides the visibility needed to understand Claude Code adoption patterns across your organization. The platform integrates data from over 100 development tools including GitHub, Jira, CircleCI, and PagerDuty, creating a unified view of how developers interact with AI coding assistants.
For Claude Code specifically, Faros AI tracks usage and adoption across three key dimensions:

Teams with declining or consistently low acceptance rates signal a need for targeted enablement efforts. The platform helps you identify these struggling teams early so you can intervene with training or support.
Power users emerge in the data. Faros AI classifies users based on frequency and consistency of usage over time. Power users demonstrate at least 20 usage days per month or activity across 50 different hours monthly. These developers achieve measurably higher output and become potential champions for broader adoption.
Faros AI enables cohort tracking for how to measure developer productivity improvements. For instance, it:
Let's understand this with a practical example. Consider an engineering director. She sees Team A at 5% Claude Code adoption versus Team B at 60%. Faros AI data shows Team B merges 47% more pull requests daily but has 35% longer review times.
What does this reveal? Team B is writing code faster with Claude Code, but the increased volume is creating a new bottleneck in code review. The team that appeared more productive based on code generation alone is actually slowing down overall delivery because reviewers can't keep pace. This is why measuring the entire workflow matters, not just coding speed.
Claude Code pricing operates on a token-based model with tiered limits. Something we discussed previously. But what does this actually cost your organization?
Most engineering leads can't answer that question. They know the license fees but not the cost of AI-generated code per unit of output.
Faros AI answers this question by measuring token usage to specific engineering outputs. The solution tracks Claude Code token consumption by model type and correlates it with commits and pull requests. This creates visibility into the true cost per commit and cost per PR across your teams.

Here's a practical example. A team of 50 developers on Max plans costs $120,000 annually. Faros AI tracks 8,400 pull requests merged versus a baseline of 5,200. The cost per incremental PR is $37.50. If each PR saves two hours of developer time at $75 per hour, the value is $150. Your ROI is 4:1.
Note: This is a back-of-the-envelope calculation; real-world results vary significantly as PR value differs enormously across different changes.
Is Claude Code worth it? The answer depends on utilization. Faros AI identifies underused licenses where developers generate fewer than 20 PRs monthly despite Max plan access. The platform flags overspend where high token costs produce minimal output improvements or declining code quality.

Engineering efficiency metrics become actionable through this visibility. Leaders can downgrade inactive users from Max to Pro plans. They can reallocate licenses to high-value use cases like technical debt reduction or legacy modernization. They can calculate break-even points for different team sizes and usage patterns.
This is how you audit real ROI from Claude usage. Connect dollars spent to business outcomes delivered.
When measuring software productivity a lot of areas need to be considered. For instance, velocity metrics alone miss critical quality and stability signals that determine organizational delivery capacity.
Faros AI tracks velocity through pull requests merged, review time, and sprint completion rates. The platform monitors how Claude Code affects these software metrics in real time. Developers using AI coding assistants show 98% higher PR throughput, but does this translate to faster delivery?
The answer requires examining what happens to these PRs and how fast they can be deployed. Code quality plays an outsized role in this. Faros AI measures test coverage percentage; bugs per developer, team, and application; and PR size distribution. Research shows AI-generated code tends to be larger and contains more defects. This concludes that we are still far away from fully trusting AI. This inherently creates downstream bottlenecks in code review and quality assurance processes.
To mitigate this issue we can consider DORA. DORA metrics provide the framework for understanding organizational impact. The four key measurements are:
Elite performers deploy on demand with lead times under one hour and change failure rates below 15%.
But how does Claude Code affect these DORA metrics? The evidence from Faros AI research shows mixed results. Despite individual developers completing 21% more tasks and merging 98% more pull requests, organizational DORA metrics remain largely unchanged. The productivity gains at the individual level don't translate to improvements in deployment frequency, lead time, change failure rate, or mean time to restore.
This is the AI Productivity Paradox. Individual output increases dramatically, but organizational delivery velocity stays flat. This is because, now, the bottleneck is shifting from code generation to code review and validation. Teams merge more PRs, but review times increase by 91%, creating new constraints downstream. The result is more code in the pipeline without faster delivery to production.
In order to resolve such issues, teams must invest in better testing infrastructure and review processes to capture productivity gains. The goal isn't just to write more code faster. It is to ensure that the increased individual output translates to faster organizational delivery without sacrificing quality.
Measurement enables optimization. Faros AI transforms Claude Code adoption from experimentation into systematic improvement of engineering productivity.
Start by identifying lagging teams in your Faros AI dashboards.
How can team leads use Faros AI reports effectively?
Review adoption patterns weekly to spot disengagement early. Compare individual token costs against output metrics to identify negative ROI users. Monitor change failure rates to catch quality degradation before it compounds. Use dashboard data in one-on-one coaching to demonstrate impact and refine usage patterns.
Developer productivity measurement tools like Faros AI enable responsible AI investment decisions. You can reallocate licenses from low-value to high-value use cases. You can justify budget increases with concrete ROI data. You can increase developer productivity systematically rather than hoping adoption equals improvement.
Observability is critical for responsible AI tooling at scale. Without measurement, you're flying blind with six-figure investments and unclear returns.
Claude Code delivers real productivity gains at the individual developer level. The organizational impact remains unclear without proper measurement infrastructure in place.
Faros AI solves this visibility problem by connecting Claude Code usage to business outcomes. The platform tracks adoption patterns, calculates cost per engineering output, and reveals where AI accelerates delivery versus where it creates new bottlenecks. Teams can prove ROI with data rather than relying on developer sentiment alone.
The evidence shows mixed results across organizational KPIs. Coding time decreases while review time increases. Pull request volume jumps but bugs per developer increase. These trade-offs require active management through quality gates and scaled review capacity.
Start with baseline measurements before rolling out Claude Code to additional teams. Run controlled experiments for at least one quarter. Track both velocity and quality metrics simultaneously. Use Faros AI dashboards to identify optimization opportunities and reallocate licenses to high-value use cases.
Responsible AI tool investment requires observability. Measure to optimize. Optimize to prove value. Prove value to scale confidently. Reach out for a demo to learn more.




