How to measure engineering productivity in 2026

The industry's most in-depth guide to measuring engineering productivity: what to track, how to collect data, and how to turn metrics into business impact at scale.

red background, illustration of book with arrows on cover

How to measure engineering productivity in 2026

The industry's most in-depth guide to measuring engineering productivity: what to track, how to collect data, and how to turn metrics into business impact at scale.

red background, illustration of book with arrows on cover
Chapters

Measuring engineering productivity in 2026

Engineering is an increasingly important and expensive function. According to BCG, the fastest-growing companies are spending more than 20% of their revenues on R&D and as much as 40% to 50% when trying to expand beyond their core products.

Engineering leaders are being asked to know their business thoroughly and explain what engineering is doing, how it relates to key company initiatives, and what resources they need and where. To say it simply, they must be able to conduct business-oriented conversations about very technical things.

Measuring engineering productivity is how leaders meet that expectation. This guide walks through what to measure, how to collect the data, how to make sense of it, and how to turn it into decisions that move the needle on business outcomes.

What is engineering productivity?

Engineering productivity is a measure of how efficiently an engineering organization delivers high-quality, functional software. Engineering productivity should be thought of as a multi-dimensional concept and not as a single number; speed of delivery is part of the equation, but so are quality, collaboration, developer experience, resource utilization, and whether the team is building the right things in the first place. Optimizing for any one of these in isolation tends to degrade the others, which is why experienced engineering leaders treat measuring engineering productivity as a balance of competing factors rather than a metric to maximize.

Why measure engineering productivity?

For a small shop with twenty engineers, leaders can usually answer questions about what's getting done by walking around. As an organization scales, that line of sight disappears. A VP of Engineering at a 500-person organization cannot directly observe how time is being spent across thirty teams, which projects are slipping, where the bottlenecks are, or how each sub-org is performing relative to the others.

Engineering productivity programs exist to restore that line of sight at scale. They give senior leaders the visibility they need to make resource and strategy decisions, and they give line managers the data to coach teams and remove blockers. When the program is working, productivity metrics inform quarterly planning, budget allocation, talent reviews, and board reporting — not just engineering's internal retros.

How to measure engineering productivity in five steps

There is not a universal "correct way" to measure engineering productivity, because what and how you measure depends on your context: your engineering productivity program should be adapted to your context, including what you need to achieve, how you work, and what you value.

Your context will naturally change over time as you grow, evolve, and respond to market forces, and your program will evolve along with these changes. To create an engineering productivity program around this context, engineering organizations should follow these five steps:

  1. Decide what to measure. Identify the engineering productivity metrics that match your goals, operating model, and culture.
  2. Collect the data. Connect the systems where engineering work happens: task management, source control, CI/CD, HR, surveys, and business systems.
  3. Normalize and validate the data. Reconcile differences across teams, tools, and workflows without forcing premature standardization.
  4. Analyze the data. Apply role-based lenses, benchmarks, and AI-assisted analysis to extract insight from telemetry.
  5. Operationalize the data. Embed engineering productivity metrics into recurring cadences — planning, retros, talent reviews, QBRs — to support true data-driven engineering.

In order to measure engineering productivity in this way, engineering leaders need visibility. At enterprise scale, steps 2 through 5 are typically run on a software engineering intelligence platform (SEIP): a category of tool that unifies engineering data across systems, normalizes it, and surfaces it as operational insight. As organizational complexity grows, the capabilities of your SEIP matter. Think: modularity, customizability, and extensibility needed to support enterprise realities.

The rest of this guide will help you identify the right path for you today and navigate the five main steps of a successful data-driven engineering productivity program.

What are the most important engineering productivity metrics?

The most useful framework for measuring engineering productivity in 2026 is SPACE. SPACE is a multi-dimensional model proposed by researchers at Microsoft, GitHub, and the University of Victoria, and it has become the dominant academic and industry reference for productivity measurement. SPACE stands for five dimensions: Satisfaction and well-being, Performance, Activity, Communication and collaboration, and Efficiency and flow.

SPACE is useful because it is holistic without being prescriptive. It combines system-generated telemetry with developer sentiment from surveys and interviews, and it explicitly recommends measuring multiple dimensions at once so that gains in one area aren't masking losses in another. Measuring multiple dimensions also makes metric gaming much harder.

If you're more familiar with DORA metrics — lead time for changes, deployment frequency, change failure rate, time to restore service, and rework rate — those are a subset of SPACE. DORA captures performance and a slice of efficiency, but it does not address satisfaction, activity, or collaboration. DORA is a strong starting point, not a complete framework.

The challenge with SPACE in practice is that it offers a wide menu of possible metrics in each dimension. Choosing the right ones requires a clear view of three things: your goals, your operating model, and your engineering culture.

How do you choose the right engineering productivity metrics?

To decide which engineering productivity metrics you should be measuring, first take a beat to identify what's important to you, how you define success, and what productivity looks like to you. In selecting what to measure, consider three elements:

  • What goals you need to achieve: Companies have different goals, priorities, and challenges at different stages of their lifecycle. A scaling ecommerce startup that just raised its Series C is optimizing for different results than a publicly traded enterprise in a highly regulated domain. This guide will help you identify what to measure at your current stage.
  • How you work: Your operating model introduces different dimensions to how you analyze and examine productivity. Do you insource or outsource? Are you hybrid or fully remote? Are you structured with regional hubs or geo-concentrated? Do you have multiple SDLCs?
  • Your engineering culture: The organization's values and DNA determine which data is collected. For example, a company that uses ranking in its formal performance management processes will likely be comfortable with individual metrics, while others may prefer to go down to the team level only.

The next three subsections walk through how each factor shapes which metrics to pick. Larger organizations will often have multiple internal groups facing different versions of these questions, in which case corporate strategy can guide where to focus first.

What engineering metrics matter at each stage of growth?

Below are typical goals and engineering productivity metrics based on the stage of your company. Metrics are additive as a company progresses through the stages.

Company Stage Engineering Goal Engineering Productivity Metrics to Track
Startup Rapid product innovation.
Remove blockers to launching new features and finding product-market fit.
• Lead time and cycle times
• Throughput
• Deployment frequency
• % delivered vs. committed
• Bottlenecks
Growth / Expansion Mature the product offering.
Develop a technical strategy to support an expanding tech stack and team. Introduce scaffolding for more teams, a larger codebase, and dependencies.
All of the above, plus:
• Production stability (uptime, MTTR)
• Platform/infra effectiveness
• Cross-team dependencies
• Code quality (coverage, test stability, smells, security)
• Team productivity comparisons and benchmarks
• Team composition
• Developer satisfaction
• Revenue per R&D FTE
Scale-up Customer growth and retention.
Balance speed with quality, safety, and reliability to support a growing customer base.
All of the above, plus:
• On-time roadmap delivery
• Velocity and quality benchmarks (DORA 5)
• SLO and SLA compliance
• Security and privacy compliance
• Resource allocation vs. ideal targets
• Onboarding effectiveness
• Developer wait time (Git performance, build time, CI reliability, test flakiness)
• R&D OpEx per R&D FTE
Maturity Maximize profitability.
Reduce costs, standardize to industry standards, and improve retention.
All of the above, plus:
• Initiative tracking and impact
• Infrastructure costs
• Migration and consolidation impact metrics
• Individual and team performance
• Talent and skill composition
Engineering productivity metrics across company growth stages

How do you measure productivity for remote, distributed, or outsourced teams?

With most modern enterprises working on a global scale, engineering teams can be heavily outsourced, geographically distributed, and remote/hybrid, and may have either a centralized SDLC or multiple SDLCs. Your company's specific operating model determines the additional lenses through which you'll want to analyze your engineering productivity metrics.

Operating Model Description Engineering Productivity Metrics to Track
Heavily outsourced The organization relies on subcontractors, usually from more than one vendor. Metrics should compare insourced to outsourced productivity, measure the value received from each vendor, and ensure institutional knowledge is captured to prevent vendor lock-in. Productivity metrics per contract type and vendor
  • Productivity per dollar spent
  • Activity per dollar spent
  • Time spent vs. target hours
  • Velocity and throughput
  • Lead time and cycle times
  • Active vs. waiting times with attention to handoffs and approvals
  • Quality of delivery (e.g., bugs per task)
  • Code, test, and documentation coverage
  • Task and PR hygiene
Geographically distributed The organization has globally distributed development centers. Metrics should assess relative productivity of each location and identify collaboration challenges. Productivity metrics per location
  • Productivity per dollar spent per location
  • Impact of cross-geo collaboration on velocity, throughput, and quality
  • Impact of cross-geo collaboration on MTTR and SLAs
Remote / Hybrid The organization has multiple employment types — in-person, hybrid, and remote. Metrics should assess relative productivity of each type, measure the impact of work-from-home policies, and inform policy, hiring, and promotion decisions. Productivity metrics per employment type
  • Onboarding effectiveness per employment type
  • Before/after impact of WFH policy changes
  • Developer experience and satisfaction per employment type
Centralized SDLC Often characterized by a monorepo. The centralized SDLC has specific impacts on developer experience that need to be tracked. Metrics should identify technical areas for optimization and pinpoint where dependencies are slowing developers down. Productivity metrics per application or service
  • PR review SLOs
  • Commit queue SLOs
  • Remote build execution and cache SLOs
  • Clean vs. cached build volume and runtimes
  • Test selection efficacy based on compute resources and change failure rate
Multiple SDLCs The organization has multiple SDLCs, often a result of a large portfolio and M&A. Metrics should identify high-performing SDLCs to cross-pollinate best practices and reduce duplication of effort. Productivity and experience comparison per SDLC
Engineering productivity metrics to track for various operating models

How does company culture affect engineering productivity metrics?

Corporate and engineering culture will also influence the smallest unit of measurement, whether the individual or the team level, and how the engineering productivity metrics are applied.

Engineering Culture Description Engineering Productivity Metrics to Track
Compete culture Employees are evaluated against strict performance metrics and stack-ranked periodically.
  • Productivity by level and against forced distribution
  • Productivity vs. ideal (expectations)
Family culture Engineering is encouraged to push toward a collective goal.
  • Productivity at the team level (not the individual)
Ad hoc culture New teams are frequently spun up to collaborate on shorter-term projects.
  • Productivity by collaboration unit (project, application, squad) as defined by GitHub teams or Jira boards
Decentralized culture Self-managed autonomous teams are given a budget and held accountable for delivering business results.
  • P&L metrics
Engineering productivity metrics to track based on company culture

Once you've identified the common engineering productivity metrics associated with your goals, operating model, and culture, you'll then need to collect the relevant data.

Where does engineering productivity data come from?

Collecting engineering productivity data is the crucial first step to creating visibility, but doing so is more challenging than it seems. The single biggest obstacle to measuring engineering productivity at scale is data fragmentation. Engineering data lives in dozens of systems — Jira, GitHub, Jenkins, SonarQube, PagerDuty, Workday, Salesforce, Google Calendar, custom internal tools — and each tool tells a partial story. Cross-tool and cross-domain analysis usually means exporting data into spreadsheets and manually stitching it together, which is slow, brittle, and unscalable.

By using an enterprise-grade SEIP like Faros, your metrics will be generated from a rich and complex combination of data sources, some standard and some bespoke, covering SaaS products and homegrown solutions, org structure data from HR systems, developer experience surveys, cost data from business systems, and much, much more.

Faros recommends a field-proven strategy for collecting engineering productivity data incrementally, creating valuable insight into productivity at each stage.

What sources should we extract engineering data from, and where do we start?

A step-wise approach can deliver quick wins, build trust, and gradually develop the data-driven mindset you need. Layer by layer, you will assemble the complete picture of engineering productivity as you've envisioned it:

Step 1: Baseline. Set aside concerns about data quality and data hygiene. Normalization and validation address those (we'll talk about this more in the next step) and are not a barrier to collection. The first step involves baselining the current state in support of your first or primary use case.

  • Most organizations start with basic questions about developer productivity: How are engineers spending their time, how long do cycles take, and where are the bottlenecks? To that end, connect to data sources for tasks and PRs.
  • To support how you want to slice and dice the information based on your operating model, you'll also need to collect the relevant metadata about your reporting structure (and occasionally, your product or app hierarchies). Connect HR data like Workday or import the data from an internal metadata service.

Step 2: Blend. Developer surveys capture developers' perceptions of how their team delivers. They provide insights into points of friction in the software delivery process and more descriptive feedback about what can be improved at the team or organizational level. Developer surveys are key to tracking employee engagement and satisfaction with the developer experience over time.

With your intelligence platform in place, the detailed and highly contextual feedback from developers can be lined up against the data you've collected from engineering systems and processes. Powerful insights come from blending qualitative insights from surveys with telemetry about systems, processes, and workflows. You'll also discover the next set of high-priority data sources to connect to for deeper analysis.

  • Collect the results of developer surveys onto the platform, such that sentiment data from employee responses can be intersected with telemetry-based data from engineering systems. This helps corroborate specific complaints about speed, complexity, and dependencies with supporting systems data.
  • Developers frequently complain that they spend too much time in meetings. Collect information about meeting load from collaboration systems like Calendars.

Step 3: Expand. Don't lose sight of the KPIs that act as checks and balances within your initiative. According to the authors of the SPACE framework, "productivity cannot be reduced to a single dimension… Only by examining a constellation of metrics in tension can we understand and influence developer productivity."

That's why, at this stage, you'll want to expand to data sources related to quality and reliability to prevent over-focusing on velocity. You've likely identified several of these metrics earlier, and you're now ready to generate them now that the basics are in place.

  • Collect data to form a holistic view of quality, reliability, and security. Pre-production sources provide a view into deployments, bugs, code smells, vulnerabilities, code and test coverage, test stability, and flakiness. Post-production sources are the systems of record for customer issues and tickets, production incidents, and system outages.
  • Collect service-level metrics to correlate "code production" metrics with the stability and efficiency metrics of supporting systems and platforms. This includes metrics like CI reliability, build cache hit ratio, development environment usage statistics, and more.

Step 4: Align. The C-Suite expects the engineering department, like every other corporate function, to demonstrate its impact on corporate objectives. To that end, the next step is collecting business results data in support of quarterly and annual planning and OKR tracking.

Some of that information may be readily available in the task management systems you've connected, which will allow you to measure say/do ratios and on-time delivery of product roadmap.

That said, there is a lot more opportunity to illuminate costs and impacts by intersecting engineering data with business metrics pertaining to product usage, customer satisfaction, and financial performance.

  • Collect data from systems that record targeted business results, like cloud costs, uptime stats, customer satisfaction, and efficiency stats for business processes supported by engineering.
  • For engineering teams that operate in lower layers of the stack, consider proxy metrics that represent value to those who consume their output.

The following table summarizes the data sources to connect at each stage of your program, with examples of specific tools.

Collection Stage Data Source Type Example Tools
Start
  • Task management
  • Source control
  • CI/CD build and deploy events
  • HR data
  • Jira, Asana, ADO
  • GitHub, BitBucket
  • Workday
Blend
  • Survey tools and spreadsheets
  • Calendars
  • Google Sheets, Airtable, Qualtrics
  • Google Calendar
Expand
  • Code quality tools
  • CI/CD build and deploy events
  • CI/CD individual test events
  • Incident management
  • SonarQube
  • Jenkins, CircleCI, Spinnaker, ArgoCD
  • ServiceNow, PagerDuty, StatusPage
Align
  • Financial systems
  • Customer experience
  • Product analytics
  • Salesforce
  • Gainsight
  • Amplitude
Engineering productivity data collection sources and example tools by stage

How do you actually collect engineering productivity data from different sources?

Standard and Custom Data Sources

During the data collection phase, the data will be stored in the Faros canonical schema, where it is normalized into a single, connected data set that can be queried efficiently across the entire organization.

Inevitably, your data sources will be a combination of vendor tools and homegrown sources, handled as follows:

  • Popular vendor tools. Faros leverages optimized open-source connectors for common task management tools like Jira and Asana, source control tools like GitHub and GitLab, static code analysis tools like SonarQube and Codacy, and others. Connectors typically pull historical data from the source once and then pull dataset changes on a cadence. When possible, we recommend also using Webhooks to push data to Faros in real-time.
  • Custom homegrown tools and sources (including spreadsheets). Two easy options exist to ingest data from custom sources.
    • Connector Development Kit: Faros leverages an open-source framework called Airbyte, which has connectors to common sources like S3 and Postgres to ease development. You'll simply have to map the data from the custom source to the schema.
    • Events CLI: For builds and deployments, it is easy to instrument the tool and have it send events to Faros. This is accomplished by adding a line of code to a preexisting CI/CD script.

Sources for Reporting Structure

Engineering productivity data is often examined based on reporting structure, with leaders tracking metrics for the sub-org and teams they manage. This essential information can be ingested from an HR source like Workday, which will update Faros when people join or leave the company or upon a major re-org.

Within Faros, the reporting structure is used for rollups and drill-downs because it represents how your teams, teams of teams, and groups are organized. It also enables comparisons, outlier identification, and team-tailored insights.

Sources for Data Attribution and Ownership

In addition to the formal reporting hierarchy, Faros infers the mapping between teams, apps, and repos. (While rare, if there is a source of truth for this data, it can be ingested directly instead.)

Furthermore, for any specific metric, your organization can choose whether to attribute it to a team based on:

  • the people whose data contributed to the work, or
  • the specific app or repo that the data touches

Faros will auto-select the best attribution method based on experience and domain expertise, but as with any other metric behavior, this is configurable.

Extending the Schema for Special Use Cases

The Faros schema is designed to represent all the relevant SDLC data in a cohesive, interconnected manner. To use it, you simply need to plug in your data.

The standard schema will be sufficient in 99% of cases. In rare edge cases where a certain data set does not fit perfectly within the schema, the schema can be extended by leveraging tags, custom columns, and custom metrics. If your use case is of general interest, Faros will consider making it a first-class concept in the schema.

How do you measure engineering productivity across teams that use different tools?

Every sizable engineering organization inevitably has different teams working in different ways. They have different workflows running on different tools and pipelines, use different custom fields and statuses, and release on different cadences.

To analyze all this data through a single lens requires normalization, and to trust the data requires confidence in its quality. But how do you trust the numbers if there's bad data hygiene?

To put it bluntly, every organization has data hygiene issues when it comes to human-curated data. That's the bad news. But here's the good news:

  • Your organization has a lot of clean machine-generated data about PRs, builds, and deployments, where data hygiene is not an issue.
  • Faros handles normalization to baseline performance and highlight hotspots, even with less-than-perfect hygiene for human-curated data.
  • In the early days of your engineering productivity program, you'll be focused on coarse, high-level metrics that baseline and benchmark performance and highlight areas of concern. This is similar to the output of an audit or a consultancy engagement. At this level, the law of large numbers makes hygiene less of an issue.

The guiding principle here is to start measuring to create high-level visibility and then let the teams identify and address pressing data hygiene issues once they see how they are skewing their metrics. Never ask teams to change how they manage projects or use Jira before they start actually generating metrics and insights. Instead, let Faros highlight the inconsistencies that teams should address.

How do you improve engineering data hygiene?

Good data quality results from the organization's commitment to becoming data-driven. Normally, no one is incentivized to address these issues until they start impacting highly visible metrics. That is to say, once leaders start paying attention to metrics, the data hygiene issues will be fixed — but not before.

The road to improving data quality involves top-down and bottom-up motions:

  • Top-down: Leaders hold their organization accountable for important metrics.
  • Bottom-up: Teams review their metrics and address any underlying data quality issues.

If you treat data quality as a showstopper, you will never get visibility. And without visibility, you'll never address the data quality issues. You'll waste years trying to address data hygiene only to discover you were focused on fixing the wrong things. Here are some simple dos and don'ts:

Best Practices for Engineering Data Quality Common Mistakes to Avoid
Start measuring for rapid baseline visibility Don't force teams to refactor or standardize their processes
Ask teams to validate the metrics and fix glaring data hygiene issues (e.g., closing tickets when work is done) Don't precondition visibility on data hygiene
Offer a small set of standard processes where it would drastically simplify measurement Don't mandate a single way of working for everyone
Best practices and common pitfalls in improving engineering data quality

How does Faros normalize engineering data across teams and tools?

In this section, we'll explain how Faros handles sophisticated normalization for common scenarios without requiring upfront standardization. This includes:

  • Multiple data sources for the same data
  • Same tool, different workflows
  • Same tool, different usage
  • Different goals for different teams
  • Basic data validation

Multiple Data Sources for the Same Data

It is quite common for different teams or sub-orgs to use different tools. For example, one team might manage their tasks in Jira, another uses Asana, and a third uses GitLab. Another example is a company with multiple instances of the same tool.

Normalization is very simple in these cases. The Faros connectors normalize the data upon ingestion, automatically mapping corresponding data types to the right place in our canonical schema.

Same Tool, Different Workflows

Another common scenario is projects within a single tool, like Jira, that use different workflows, as expressed by statuses. Faros automatically deals with status transitions and provides the desired breakdowns based on the level of analysis:

  • Each team's particular workflow will be represented in its metrics, so team members can understand their bottlenecks, learn, and affect change where needed.
  • At the leadership level, where we're zoomed out to team-of-teams or much larger groups, metrics will be abstracted to the common statuses of To Do, In Progress, and Done. This is sufficient to see the bottom line metrics leaders care about, like task cycle time and amount of work in progress.

Same Tool, Different Usage

Every team in your organization might be using Jira, but they're using it very differently. Normalization is required to report effectively across this variance in tool usage.

The Faros approach is to be compatible with how people work today, especially at the very beginning of your program. To that end, the data normalization can be handled in a couple of ways:

  • By building conditions into the chart queries. For example, let's imagine you want to look at all high-priority unassigned issues. One team may use P0 and P1, while another uses Critical and High. A custom query can bake these different definitions into the chart.
  • By using the platform's data transform capabilities. For example, one group uses epics to track initiatives, while another group uses tags on tasks, and a third group uses a custom issue type. Faros transforms the data to the initiative portion of our schema, so you can then query all the initiatives in a single way.

At some point, if the maintenance of the queries or transforms becomes too complex and error-prone, Faros recommends introducing a few standard options. You don't force everyone to comply to the same behavior, rather to select one of a handful of approved ways of doing things. This should cover the majority of team preferences while keeping the in-tool configurations manageable.

Different Goals for Different Teams

Let's face it: good and bad are relative.

Consider one product under active development and another product that is in maintenance mode. While you may want to measure the same things for these teams — for example, throughput — they will have different definitions of good. "Good" is also relative to a baseline, and their starting points may be wildly different.

The Faros approach is to make it easy for every role to easily understand how teams are performing relative to contextual goals.

  • Teams can customize their thresholds for great, good, medium, and bad. These custom thresholds will be utilized for their personalized dashboards featuring team-level metrics and insights.
  • Leaders will get a bird's-eye view at the organizational level that takes all the personalized thresholds into account and visually identifies hotspots. It will also call out areas of improvement or decline.

Note: Popular frameworks like DORA publish annual benchmarks, but the way the metric is defined might not be applicable to how you work. For example, deployment frequency measures how often you deploy code changes to production. If your organization has a major product release four times a year, strict adherence to that definition won't give you the insight you seek. In this example, Faros recommends measuring deployment frequency to your pre-prod environments.

Basic Data Validation

As Faros begins to ingest and normalize data, it will identify gaps in data collection and mapping. Through troubleshooting and cleanup, you can address and fix these errors. In addition, charts can be tweaked, for example, to use median instead of average.

During this phase, you may also discover places where processes are not being followed internally and need to address the issue with the relevant teams.

  • Faros produces a report that shows data gaps, including boards without an owner and teams without data.
  • Faros highlights anomalies and outliers, primarily found in human-curated data, like tasks that have been open for hundreds of days or tasks that moved directly from ToDo to Done.

Once your data has been validated, you've paved the way to the next stage of analyzing it.

What do you do with engineering productivity data?

Once data is flowing and normalized, dashboards and scorecards begin to populate. The next question is what to do with them. Analysis is where measurement either pays off — by revealing contributing factors and pointing to changes — or fails, by drowning leaders in dashboards no one acts on.

The most useful frame for analysis is role-based. Different roles in the organization care about different questions, at different levels of aggregation, with different cuts of the data. A senior engineering leader looking at organizational health needs a different view than a line manager debugging a sprint, who needs a different view than a TPM tracking a cross-functional initiative.

Engineering productivity data exists to serve two distinct purposes:

  1. To answer questions you already know to ask. "How are we doing on the five reliability KPIs we're prioritizing this year?" Dashboards and scorecards handle this.
  2. To surface things you wouldn't have thought to ask. "The Mobile and Data Platform groups have engineering manager-to-developer ratios well above industry benchmarks, suggesting an opportunity to rebalance." These insights come from curated analytics and AI-driven analysis layered on top of the raw data.

Which engineering productivity metrics matter most for each role?

The table below maps engineering roles to their analysis dimensions and the questions they should answer first when a measurement program goes live.

Engineering Role Recommended View Questions to Answer First
Senior Engineering Leaders (VP, CTO, Director) Top-down for the organization, split by sub-org and team
  • KPIs, overall health, and hotspots
  • Key initiative progress
  • Cost vs. impact
  • Resource allocation
Line Manager Bottom-up for the team, split by service, repo, and IC
  • Baseline velocity and quality
  • Performance trends
  • Bottlenecks and dependencies
Functional Leaders (Product, TPM, PPM) Horizontal for the domain, split by team, epic, initiative
  • On-time delivery
  • Process efficiency
  • Planning efficacy
  • Coaching efficacy
  • Bottlenecks and dependencies
Functional Leaders (Platform Engineering, DevEx) Horizontal for the domain, split by team, repo, and pipeline
  • Tool utilization and impact
  • Delivery velocity and bottlenecks
  • Developer satisfaction
Functional Leaders (QA, Security) Horizontal for the domain, split by team, repo, or service
  • KPIs, SLAs, SLOs
  • Cost vs. impact
Initiative Leaders (Migration, Modernization, AI) Horizontal for the initiative, split by team
  • Resource utilization
  • Budget planned vs. actuals
  • Risk
HR Leaders Top-down for employees, split by geo, employment type, tenure, manager
  • Engagement and retention health
  • Skill and composition health
  • Hotspots
Engineering productivity metrics and views by role

The day-one goal for any role is the same: get a sense of overall health, confirm or correct gut feelings, and identify hotspots. Deeper analysis comes after.

How do you validate and contextualize engineering productivity insights?

Now that you've baselined your data, identified hotspots, bottlenecks, or areas of friction, and begun uncovering their contributing factors, it's time to validate and contextualize the information.

No one knows your business like your developers and managers, so any statistical finding should be validated by the people involved. Faros recommends the following continuous improvement approach with four stages: Measure, Understand, Decide, Act.

  1. Measure: Gather the data points indicative of an issue.
  2. Understand: Deepen your understanding of the issue by adding context. 
    • Talk to your team. Speak to your team members to get their perspective. Learn what they think is happening and why. This is where you tie the telemetry to all your institutional knowledge and business context. 
    • Consult developer survey data. Developer surveys should ask developers what's causing the most friction in their daily work. If you're not yet importing your qualitative developer survey data into Faros to correlate it with quantitative data, consider doing that now. If not possible, manually review recent developer surveys to see what they say about the issue.
    • Confirm the priority. How important and consequential is this issue? Arrive at a consensus on whether this issue is worth pursuing now.
  3. Decide: Formulate a hypothesis and customize visibility to accompany your analysis. 
    • Formulate a hypothesis, time-box it, set a goal, and consider the different levers you have to take action. 
    • Set up granular metrics for the workflow or process in question, if required. For example, if the issue pertains to incident resolution, configure a flow in Faros that will allow you to examine every step of the process across tools and interactions like Detect > Create > Triage > Resolve > Restore. This will help you measure the impact of your changes throughout its cycles. 
    • Create new charts or dashboards that will enable you to examine the factors at play. For example, if your hypothesis is that incident resolution times are influenced by cross-geo delays, customize a chart to incorporate the geographical location of the team members involved. 
    • Configure alerts or notifications that will be helpful in tracking or addressing the issue.
  4. Act. 
    • Enact the change. Designate someone to be focused on the issue and the change management process. They should enable the teams, monitor the impact, and share the learnings. 
    • Monitor the impact. After acting, keep your eyes on the metrics. Is the change helping you achieve your goal? Are there unintended consequences you need to address?

How do you make engineering productivity data useful?

The most underestimated step in engineering productivity programs is operationalization. Tools don't change companies; the use of tools in recurring decision-making does. Implementing a software engineering intelligence platform is similar to implementing Salesforce. Salesforce doesn't increase sales by being installed; it increases sales because the Head of Sales reviews metrics weekly, salespeople keep their pipelines current, QBRs are run on Salesforce data, and decisions are visibly tied to the numbers.

Engineering needs the same discipline. Without it, a measurement program produces dashboards no one looks at and reports no one references in meetings. With it, engineering metrics start showing up in planning sessions, retros, talent reviews, and board reports, and decisions get faster and more confident as a result.

For most organizations, getting there requires change management: modifying existing meeting protocols and practices to include integrated data. Four guiding principles support that transition.

Change Management Principle What to Actually Do
Establish ownership and accountability for change Designate an internal champion with the authority to institute new data-driven practices
Tailor visibility to support recurring cadences Build customized dashboards for each meeting type so the data is ready for the conversation
Make resource allocation and decisions contingent on supporting data Require data evidence for resource, prioritization, and HR-related decisions
Make every team accountable for its own data and improvement Train teams on their metrics and on the steps for ensuring data quality
Change management principles and recommended actions

What are the five pillars of engineering operations?

World-class engineering organizations, from scaling startups to mega-enterprises, run on five operational pillars: Productivity, Delivery, Outcomes, Budgets, and Talent. Each pillar is supported by recurring meetings and decision processes, and each benefits from being fueled by integrated data instead of partial spreadsheets and gut feel.

Most organizations launch their measurement programs with the productivity pillar — that's where engineering's internal pain usually lives — and then expand outward to the other four. The remaining sections walk through the cadences and recommended metrics for each pillar.

Productivity cadences and metrics

Platform Engineering, Developer Experience, and Architecture teams run continuous initiatives to modernize technology, optimize workflows, and remove friction. Monthly operational reviews track key metrics, address challenges, and align on priorities. Project reviews track foundational transformations like migrations, modernization efforts, compliance initiatives, and tooling rollouts.

Productivity Cadence Engineering Productivity Metrics to Review
Group-level operational review
  • On-time release success rates
  • Release delays vs. target
  • Feature rollout progress
  • Recent service-level metrics
  • Production incident review (SEVs)
  • Tool/process adoption
Product or team technical review
(analyzed for each component within the team's scope of ownership or from the product lens)
  • Deployment frequency
  • What shipped this quarter
  • Activity volume by work type
  • Velocity
  • Service-level metrics (uptime, performance, production incidents, KTLO, production defects, vulnerabilities by severity)
  • Qualitative feedback
Periodic developer experience survey review
  • Analysis of qualitative responses
  • Correlation of qualitative responses to quantitative data
  • Hotspot identification
Productivity cadence and metrics to review

Delivery cadences and metrics

Delivery in most organizations runs on agile, with work segmented into sprints or development iterations. Stand-ups, sprint planning, and retros keep projects on track and quality high. Monthly product and tech reviews provide a structured forum for assessing progress, surfacing issues, and aligning priorities.

Delivery Cadence Engineering Delivery Metrics to Review
Quarterly planning Previous quarter look-back to inform next quarter:
  • Velocity, throughput, and quality metrics for capacity estimation, prioritization, and commitment-setting
  • Resource allocation across initiatives and work types
  • Previous quarter's work distribution
  • % completion of initiatives and analysis of outstanding work
  • Cross-team dependencies and bottlenecks
  • Adherence to targeted resource allocation (strategic/innovation, bug fixes, KTLO, tech debt)
Sprint retros Planning accuracy and predictability:
  • Say/do ratios
  • Unplanned work percentage
  • Developer satisfaction
  • Task and PR hygiene
  • Backlog health
  • WIP
Progress vs. goals:
  • % completion of epics and analysis of outstanding work
Initiative / project reviews
  • Initiative progress vs. plan and budgets
  • Predicted delays and cost overruns
  • Breakdown of remaining work by epic and team
Delivery cadence and metrics to review

Outcomes cadences and metrics

Outcomes are managed on a quarterly rhythm — setting, reviewing, and adjusting OKRs. Regular check-ins keep teams aligned to strategic goals, and cross-functional QBRs evaluate the quarter's performance. The outcomes pillar is where engineering data has the most direct impact on the C-suite conversation, because it's where engineering metrics get tied to business results.

Outcome Cadence Engineering Outcome Metrics to Review
C-suite and board reporting
  • Engineering as a % of revenue
  • Revenue per engineer
Quarterly reviews (start, mid, end of quarter) All of the above, plus: Initiative progress:
  • Initiative progress vs. goals and risks
  • Remaining work per initiative
  • Risk mitigation strategies
  • Cost vs. impact
Business metrics:
  • Growth metrics
  • Customer retention metrics
  • QoQ North Star metrics (adoption, usage, revenue, retention, security, reliability, CSAT)
  • Engineering as a % of revenue
  • Revenue per engineer
Roadmap metrics:
  • Delivery update (completed, in progress, up-next)
  • Committed work by investment category
  • Delivered work by investment category
  • Resource allocation across the portfolio
Portfolio / product KPIs:
  • % delivered vs. committed and cause/reason
  • Initiative progress vs. goals
  • Productivity vs. goals (velocity, throughput)
  • Quality vs. goals (critical defects, open and closed bugs by team, code coverage)
  • Availability SLAs (closed bugs SLA, open bugs SLA, % bugs solved out of SLA)
  • Support SLAs (MTTR, code coverage, CSAT)
  • Security SLAs
  • Bottlenecks and dependencies
Outcome cadence metrics to review

Budget cadences and metrics

Engineering budget planning runs on an annual cycle, with quarterly reviews and adjustments. The process forecasts financial needs, allocates resources, and sets goals for the upcoming year. Vendor and global sourcing reviews happen at their own cadences, often tied to contract renewal cycles. Periodic accounting events, such as capitalization reviews, also fall in this pillar.

Budget Cadence Engineering Budget Metrics to Review
Annual planning and quarterly review Impact:
  • Engineering as a percentage of revenue
  • Revenue per engineer
  • ROI by engineering initiative
  • Resource allocation vs. revenue
Productivity:
  • Productivity vs. benchmarks
  • Productivity per location
Efficiency:
  • Organization composition by role
  • Management overhead vs. benchmarks
  • Functional staffing ratios vs. benchmarks
Global sourcing review and planning
  • Productivity per dollar spent per location
  • Productivity per employment type
  • Productivity per contract type
Vendor and subcontractor contract negotiations and renewals
  • Productivity per dollar spent per subcontractor
  • Vendor costs and ROI
Periodic accounting events
  • Capitalized software development costs per period
Budget cadence and metrics to review

Talent cadences and metrics

Talent reviews and performance evaluations typically happen twice a year. Compensation reviews run in March or April for most organizations. Talent decisions benefit from objective data more than almost any other engineering process, because objective data reduces subjectivity, accelerates preparation, and grounds feedback in evidence rather than impression.

Talent Cadence Engineering Talent Metrics to Review
Performance reviews Impact:
  • PRs authored and tasks completed with respect to size, complexity, and impact (over time, by work type, by epic)
  • PR and task cycle times relative to team and peers
Quality:
  • PR quality (size, code quality analysis, review time, comments, reverts)
Collaboration:
  • PR reviews completed
  • Authors reviewed
  • Reviewers
  • Bugs and incidents resolved
Calibration:
  • Comparison to similar cohorts by role, seniority, and tenure
Talent reviews and workforce planning
  • Number of engineers (by level, by job profile)
  • Functional staffing ratios vs. benchmarks
  • Team composition outliers vs. benchmarks
  • Open headcount
  • Attrition rate
  • At-risk attrition employees
  • Technology and coding language usage
  • Developer CSAT
Organization structuring
  • Productivity per location
  • Productivity per employment type
  • Productivity per dollar spent
  • Engineering overhead ratio
  • Cross-team dependencies
Talent cadence and metrics to review

A note on individual-level metrics: whether to use them at all is a cultural question, and not every organization should. Where individual metrics are used, they should always be looked at in cohort context — comparing engineers of similar role, seniority, and tenure rather than across the whole organization — and they should be one input among several, not the basis for ranking decisions.

The bottom line on measuring engineering productivity

Engineering productivity programs that follow this sequence — measure, collect, normalize, analyze, operationalize — produce visibility that informs real decisions. They give engineering leaders the data to talk about their organizations in the same business terms as the rest of the C-suite, and they give teams the feedback loops to improve continuously. The right approach is the one that fits the organization's goals, operating model, and culture; the wrong approach is to wait for perfect conditions before starting at all.

For enterprise organizations, Faros is the only SEIP built to handle real-world complexity. Learn more about how we can help you improve engineering productivity at scale.

Summary: how to measure engineering productivity in 2026

Measuring engineering productivity is a five-step discipline. Each step has a guiding principle and a concrete set of next actions.

Question Guiding Principle What To Do
What should we measure? Identify what matters to you.
Tailor your program to your business context, otherwise there will be no buy-in.
Identify your goals based on your company’s stage.
Understand how your operating model dictates key analysis dimensions.
Take care to preserve your engineering culture in your approach to metrics.
Which data sources should we collect? Begin with the basics and advance.
Take a step-wise approach to achieve quick wins while gradually completing the picture.
Baseline with task, PR, and org chart data.
Blend in developer surveys and collaboration tools.
Expand to form a holistic view of quality, reliability, and security.
Align to corporate objectives with business outcome tracking.
How can we normalize and validate the data? Don’t let the perfect be the enemy of good.
Data quality is a result of visibility, not a pre-condition. Start measuring to create high-level visibility.
Metrics will highlight the inconsistencies that teams should address without enforcing wholesale standardization.
If leaders demonstrate they care, teams will address the data gaps, anomalies, and outliers identified by Faros.
How do we begin to understand our current state? Take a role-based approach.
Get a sense of overall health for the performance dimensions you care about, validate gut feelings, and identify hotspots.
Gauge the current state with benchmarks.
Counteract data overload with visualizations and AI.
Validate and contextualize findings.
Perform ad-hoc, impromptu analysis to answer emerging questions.
How do we operationalize the data to achieve impact? Change management is essential to becoming a data-led organization.
Inject data into your recurring cadences and decision processes across the five core pillars of engineering operations.
Productivity: Monthly operational reviews and project reviews.
Delivery: Quarterly planning, sprint retros, and initiative reviews.
Outcomes: C-Suite and board reporting, quarterly reviews, and midpoint check-ins.
Budgets: Annual planning, quarterly reviews, vendor contract negotiations, global sourcing strategic planning, and periodic accounting/compliance events.
Talent: Performance reviews, workforce planning, organization structuring.
Guiding questions, principles, and recommended actions for engineering productivity programs

Frequently asked questions about measuring engineering productivity

What's the difference between SPACE and DORA metrics?

DORA is a five-metric subset of SPACE focused specifically on software delivery performance. The DORA metrics are lead time for changes, deployment frequency, change failure rate, time to restore service, and rework rate (added in 2024 to measure unplanned deployments addressing user-facing bugs). SPACE is the broader framework, covering five dimensions: Satisfaction and well-being, Performance, Activity, Communication and collaboration, and Efficiency and flow. Use DORA as a starting point for measuring delivery; use SPACE for measuring engineering productivity overall.

Should engineering productivity be measured at the individual or team level?

That depends on engineering culture, but we recommend sticking to the team level as the smallest unit of measure. Organizations with compete cultures and stack ranking are usually comfortable with individual metrics. Organizations with collective-ownership cultures should stay at the team level. Where individual metrics are used, they should compare engineers within similar cohorts by role, seniority, and tenure, and they should never be the sole input into compensation or performance decisions. Most enterprise engineering organizations get more value from team-level measurement than from individual-level measurement.

How long does it take to set up an engineering productivity program?

Initial baseline visibility (task data, PR data, and org structure connected and producing dashboards) usually takes weeks, not months. Reaching a comprehensive program that covers all five operational pillars (productivity, delivery, outcomes, budgets, talent) is typically a multi-quarter journey, with each stage producing its own value before the next begins.

Do you need clean data before you start measuring engineering productivity?

No, and waiting for clean data is the most common reason programs never launch. Machine-generated data from PRs, builds, and deployments is mostly clean by default. Coarse, high-level metrics are robust to hygiene issues thanks to the law of large numbers. Once leaders start paying attention to metrics, teams have a reason to clean up the underlying data, and they will. Visibility produces data quality, not the other way around.

How do you avoid data overload when measuring engineering productivity?

Two tools handle data overload in engineering productivity measurement: visualization and AI. For visualization, use simplified scorecards that consolidate the top 5–10 KPIs the organization cares about. This lets senior leaders see which areas are healthy and which require attention, with drill-down available when needed. For AI, use statistical analysis and machine learning to identify problem areas in specific sub-orgs, repos, or stages of the SDLC, and surface team-tailored insights about what's helping or hurting performance. Both reduce the time between identifying an issue and acting on it.

How do industry benchmarks help with engineering productivity?

Industry benchmarks let an organization see itself in context, answering three questions that are otherwise hard to answer in isolation: where to start, what to aim for, and how to justify investment in incremental improvement. Most popular benchmarks come from research that ties high engineering performance to better financial performance, which gives them credibility in business conversations. Common benchmarks include the DORA 5 metrics, cycle times, velocity, say/do ratios, planned vs. unplanned work, AI coding assistant impact, and staffing ratios. They are most useful for surfacing the performance areas where the gap between current state and industry norm is widest.

What metrics should a startup measure vs. a mature company?

Startups should focus on lead time, cycle times, throughput, deployment frequency, percent delivered vs. committed, and bottlenecks. These are the metrics that surface friction in shipping new features. As the company grows, additional dimensions get added: production stability and code quality during the growth stage; on-time roadmap delivery and SLO compliance during scale-up; cost, individual and team performance, and skill composition at maturity. The metrics are additive: a mature company tracks everything a startup tracks, plus more.

How do you measure engineering productivity for outsourced or vendor-led teams?

The metrics are similar to those for insourced teams (velocity, throughput, lead time, cycle time, quality), but they should be sliced by contract type and vendor. The most useful additional metrics are productivity per dollar spent, activity per dollar spent, time spent vs. target hours, and quality of delivery (bugs per task). Tracking institutional knowledge capture is also important to prevent vendor lock-in.

Can you measure deployment frequency if your organization only releases quarterly?

Yes, but the standard definition needs adjustment. Strict DORA defines deployment frequency as deploys to production, which is meaningless for an organization that ships to production four times a year. The right adaptation is to measure deployment frequency to pre-production environments such as staging, QA, and integration. Same metric, calibrated to the actual SDLC.

How do you avoid metric gaming?

Three practices reduce gaming. First, measure multiple dimensions in tension (velocity and quality, throughput and stability) so that gaming one metric shows up as degradation in another. Second, don't tie individual compensation directly to engineering metrics; the moment a metric becomes a target, it stops being a measurement. Third, focus on team-level metrics where possible, since team norms tend to self-correct against gaming behavior that an individual incentive structure would reward.

What's the difference between leading and lagging engineering productivity indicators?

Leading indicators predict future performance: PR review time, build reliability, developer satisfaction, planned vs. unplanned work. Lagging indicators report past performance: on-time delivery, customer-reported defects, mean time to restore. A balanced engineering productivity program tracks both. Leading indicators tell leaders where to act; lagging indicators tell them whether the actions worked.

How do AI coding assistants affect how engineering productivity is measured?

AI coding assistants don't change the core engineering productivity metrics, but they do change benchmarks for some of them and add new metrics specific to AI assistant usage. The underlying productivity question, whether the organization is delivering high-quality, functional software efficiently, stays the same. New AI-specific metrics worth tracking include adoption, code acceptance rate, and downstream quality of AI-assisted code. The risk to watch for is over-rotating on activity metrics like lines of code or PRs opened, which AI inflates without corresponding increases in delivered value.

Who should own the engineering productivity program?

In most organizations, the program is owned by an engineering productivity, developer experience, or platform engineering team, with executive sponsorship from a VP of Engineering or CTO. The owning team needs both technical understanding (to work with the data and tooling) and organizational standing (to drive change management across teams). At least one dedicated data analyst, deeply familiar with the business and the engineering organization, is recommended.

What's the biggest mistake organizations make when measuring engineering productivity?

The most common mistake is picking a single metric (usually velocity, story points, or lines of code) and trying to maximize it. This invariably degrades quality, satisfaction, and collaboration in ways that the chosen metric doesn't capture. The next most common mistake is delaying the program until data quality is "ready," which it never will be. Both mistakes share a root cause: treating engineering productivity as a number to optimize rather than a multi-dimensional reality to understand.

Naomi Lurie

Naomi Lurie

Naomi Lurie is Head of Product Marketing at Faros. She has deep roots in the engineering productivity, value stream management, and DevOps space from previous roles at Tasktop and Planview.

AI Is Everywhere. Impact Isn’t.
75% of engineers use AI tools—yet most organizations see no measurable performance gains.

Read the report to uncover what’s holding teams back—and how to fix it fast.
Cover of Faros AI report titled "The AI Productivity Paradox" on AI coding assistants and developer productivity.
Discover the Engineering Productivity Handbook
How to build a high-impact program that drives real results.

What to measure and why it matters.

And the 5 critical practices that turn data into impact.
Cover of "The Engineering Productivity Handbook" featuring white arrows on a red background, symbolizing growth and improvement.
Graduation cap with a tassel over a dark gradient background.
AI ENGINEERING REPORT 2026
The Acceleration 
Whiplash
The definitive data on AI's engineering impact. What's working, what's breaking, and what leaders need to do next.
  • Engineering throughput is up
  • Bugs, incidents, and rework are rising faster
  • Two years of data from 22,000 developers across 4,000 teams
Blog
5
MIN READ

AI coding ROI that services the CFO conversation

AI coding tool prices are climbing. Learn how to build a defensible ROI calculation at the tool and team level — and justify spend to your execs.

Blog
6
MIN READ

AI is making engineers faster. So why does everything feel slower?

AI tools speed up developers, but delivery feels slower. Here is what the data tells us about the gap between output and absorption.

Blog
7
MIN READ

Developer sentiment surveys made sense before AI. Now they're misleading you.

Developer sentiment surveys track how engineers feel. But when AI coding tools cost real money and your CFO wants ROI, feelings aren't the answer. Here's what is.