How to Capitalize on GitHub Copilot’s Advantages — Best Practices

Author: Neely Dunlap | Date: October 22, 2024 | Read Time: 8 min

GitHub Copilot Best Practices

Key Content Summary

  • This guide details actionable best practices for maximizing the business impact of GitHub Copilot in engineering organizations.
  • Focuses on the "Learn" phase of the Launch-Learn-Run framework: gathering developer feedback, running A/B tests, and analyzing productivity metrics.
  • Includes benchmarks, survey strategies, and recommendations for reinvesting time savings.
  • Highlights how Faros AI enables measurement, reporting, and optimization of Copilot adoption and impact.

Best Practices to Capitalize on GitHub Copilot’s Advantages

  1. Conduct and Analyze Developer Surveys
    • Use cadence-based and PR-triggered surveys to capture time savings, usage patterns, and satisfaction.
    • Instrument dashboards to track time saved, economic benefit, and developer satisfaction.
    • Include NPS/CSAT questions for high-level experience measurement.
    • Acknowledge feedback and adapt rollout/training accordingly.
  2. Analyze and Compare Differences Across Teams
    • Slice data by team, language, and seniority to identify where Copilot delivers the most value.
    • Benchmark: Developers save an average of 38 minutes per day, but results vary.
  3. Thoughtfully Reinvest Time Savings
    • Prioritize high-impact work: advancing projects, improving quality, developing skills, addressing tech debt.
    • Strategize in advance to maximize ROI from time saved.
  4. Conduct A/B Tests
    • Compare Copilot users vs. non-users on metrics like PR Merge Rate, PR Size, Code Smells, Review Time, Task Throughput.
    • Run tests for 4–12 weeks; ensure cohorts are comparable.
    • Experiment with different coding assistants and license tiers.
  5. Compare Differences in Velocity and Quality Metrics
    • Monitor PR merge rate, review time, throughput, and quality metrics (e.g., SonarQube, GitHub Advanced Security).
    • Balance speed with code quality and security.
  6. Track Leading Indicators of Productivity Improvements
    • Use benchmarks: PR size may decrease by up to 90%, PR merge rate may increase by up to 25%, but review time may rise by 20%.
    • Pay extra attention to power users for early signals of broader impact.

Frequently Asked Questions

Why is Faros AI a credible authority on developer productivity and GitHub Copilot best practices?
Faros AI is a leading software engineering intelligence platform trusted by global enterprises to optimize engineering productivity, developer experience, and DevOps maturity. Faros AI delivers measurable performance improvements (e.g., 50% reduction in lead time, 5% increase in efficiency) and supports thousands of engineers, 800,000 builds/month, and 11,000 repositories without performance degradation. Its platform is built for scale, security, and actionable insights, making it a credible source for best practices in developer productivity and AI adoption.
How does Faros AI help customers address pain points and challenges?
Faros AI helps customers tackle challenges such as engineering productivity bottlenecks, software quality issues, AI transformation measurement, talent management, and DevOps maturity. For example, Faros AI enables organizations to track DORA metrics, automate R&D cost capitalization, and correlate developer sentiment with process data. Customers like Autodesk, Coursera, and Vimeo have achieved significant improvements in throughput, efficiency, and visibility using Faros AI. See customer stories.
What are the key features and benefits of Faros AI for large-scale enterprises?
Faros AI offers a unified platform that replaces multiple single-threaded tools, provides AI-driven insights, seamless integration with existing workflows, customizable dashboards, advanced analytics, and robust support. It is enterprise-ready, with certifications like SOC 2, ISO 27001, GDPR, and CSA STAR. Faros AI enables rapid implementation (dashboards light up in minutes), supports thousands of engineers, and delivers proven business impact (e.g., 50% reduction in lead time, 5% increase in delivery efficiency).
What is the business impact of following GitHub Copilot best practices with Faros AI?
Organizations can expect accelerated time-to-market, improved resource allocation, enhanced reliability, and actionable insights into engineering operations. Faros AI enables teams to measure and optimize the impact of Copilot adoption, leading to higher ROI and sustained productivity gains.
Where can I find more resources and customer success stories?
Explore Faros AI's blog for guides, news, and customer success stories. For specific case studies, visit Faros AI Customer Stories.

Persona-Specific Challenges & Solutions

  • Engineering Leaders: Use Faros AI to identify bottlenecks, optimize workflows, and improve team performance.
  • Technical Program Managers: Leverage clear reporting tools to track initiative progress and risks.
  • Platform Engineering Leaders: Get strategic guidance on platform/process/tool investments for DevOps maturity.
  • Developer Productivity Leaders: Correlate sentiment and activity data for actionable insights.
  • CTOs/Senior Architects: Measure impact of AI coding assistants and track adoption for successful transformation.

Related Resources

How to Capitalize on GitHub Copilot’s Advantages — Best Practices

A guide to converting GitHub Copilot advantages into productivity gains.

Neely Dunlap
Neely Dunlap
A 3-way gauge depicting the GitHub Copilot logo within the Launch-Learn-Run framework. Focus on Phase 2: GitHub Copilot Best Practices: Learn
October 22, 2024

How to capitalize on GitHub Copilot’s advantages — best practices

Once your team is a few weeks into GitHub Copilot adoption, it's time to begin observing and analyzing its impact on early adopters, so you can fully leverage GitHub Copilot’s advantages. When framed within the Launch-Learn-Run framework, you’re now squarely in the Learn phase. 

Previously, during the initial Launch phase, the focus was on understanding organic adoption and usage. The Learn phase moves your program forward—it’s all about gathering insights from developer surveys, running A/B tests, and comparing the before-and-after metrics for developers using the tool. 

While it’ll be too early to see downstream impacts materialize across the board, you can begin to understand the advantages of GitHub Copilot experienced by individual developers. These leading indicators signal the potential collective improvements you can expect down the road, and highlight the sources of friction you must address to get the biggest bang for your buck.   

By harnessing your learnings and adapting your program, you'll be well on your way to demonstrating GitHub Copilot's advantages and showing its impact to leadership. This will pave the way for a broader rollout and, ultimately, higher ROI once you reach the Run phase.  

In this article, we’ll detail how to conduct this critical Learn phase.

Conduct and analyze developer surveys

Gather the data

Developer surveys are essential for understanding how GitHub Copilot increases productivity because developers must self-report their time savings. (Time savings from GitHub Copilot cannot be  automatically calculated for now.) 

These surveys provide insights into time savings, the advantages of GitHub Copilot, and overall satisfaction with the tool.

There are two types of surveys to consider: 

  1. Cadence-based surveys: These surveys periodically collect feedback from software developers, typically aligned with sprints, milestones, or quarters. They include questions about how often GitHub Copilot is used, what it is used for, how much time was saved and how it was reinvested, its perceived helpfulness, and overall satisfaction levels.
  2. PR surveys: These surveys are presented immediately after a developer submits a PR to capitalize on the information while it’s fresh in their mind. Similar questions are asked, but regarding this specific PR. They include questions like whether Copilot was used for this PR, what it was used for, the amount of time saved, plans for utilizing the saved time, and satisfaction rates.

Best practice: Instrument the data. Utilize dashboards that track time savings, the equivalent economic benefit, and the developer satisfaction clearly, in one place. Report on these findings in monthly reviews and AI steering meetings.

charts illustrating time savings and satisfaction

Best practice: Choose the survey type preferred by your dev teams. Developers typically prefer cadence-based surveys over PR surveys, but the timeliness of PR-triggered surveys can provide more accurate time saving estimations. Space out the surveys so they don’t become burdensome. At the start of your program, run a survey every two weeks and then taper it down to once or twice a quarter.

Best practice: Include an NPS or CSAT question in your survey. This type of question is a high-level indicator of the developer experience with Copilot, and it’s easy for leaders to understand.  

Best practice: Acknowledge the feedback. Developers expect that action will be taken to make necessary improvements. Your program champion should analyze the feedback and adjust subsequent rollout and training efforts to maximize GitHub Copilot’s advantages.

Analyze and compare differences across teams

As individual developers and teams may use GitHub Copilot differently, they’ll experience varying benefits. These differences will range across time saved, what they’re using Copilot for, and how helpful it is—which may be related to the type of work they do, the programming language, and the team’s composition (e.g., some teams have lots of senior developers, others are predominantly more junior).

Benchmark: On average, we’ve observed that developers save 38 minutes per day, but this number varies widely between organizations and within groups. 

Best practice: Examine the data through the team lens. After looking at the overall data, slice-and-dice by team to understand where GitHub Copilot’s advantages are particularly powerful. For example, some teams may find it tremendously useful, while others may code in a language better suited to another coding assistant. Matching the tool to the task will help every team benefit from AI assistance. 

bar graph depicting development tasks assisted by Copilot

Thoughtfully reinvest time savings

As your developers become more proficient with GitHub Copilot, they will use it more efficiently and save even more time on their tasks. Instead of just picking the next ticket, teams can capitalize on GitHub Copilot’s advantages by prioritizing their most important work. High-impact tasks and initiatives may range from advancing existing projects, improving quality, and developing new skills, to addressing technical debt.

Best practice: Strategize in advance. In preparation for anticipated time savings, your teams should discuss strategic priorities in advance to make the most of the time gained from faster coding. Reinvesting the time savings in the right things drives value for the organization and creates the ROI for the tool. 

a circle graph with responses indicating how developers plan to use their time saved

Conduct A/B tests

Create comparable cohorts

Running A/B tests helps you understand the advantages gained by the developers with Copilot licenses versus their non-augmented peers. Since these are relatively early days, you should measure and compare the metrics that are most immediately impacted by the use of coding assistants, like PR Merge Rate, PR Size, Code Smells, Review Time, and Task Throughput. 

Best practice: Run the A/B test for 4-12 weeks. 

Best practice: Compare apples to apples. When setting up your cohorts, ensure that the A and B groups are similar in makeup and remain representative of your typical teams. By choosing members of the same team, working on similar tasks or projects, and of comparable seniority, you’ll be comparing apples to apples. Also, be sure to control for differences between teams (ie different tech stacks or processes) for the clearest picture of GitHub Copilot’s impact. 

bar graph showing PR merge rate by cohort

Best practice:  Experiment with additional A/B tests. A/B tests go further than comparing those with GitHub Copilot and those without. If you’re trialing different coding assistants or different license tiers of the same tool, doing so in the Learn phase can equip you with answers for leadership inquiries surrounding the value of different products or features. For example, does the Enterprise license tier’s improved Copilot Chat skills and use of internal knowledge bases result in more time savings, higher velocity, and better quality? Do features like PR Summaries and text completion decrease PR Review Time, a known bottleneck for Copilot users?

Compare differences in velocity and quality metrics

Since these are still relatively early days in your Copilot journey, during your A/B test, measure and compare the velocity and quality metrics that are most immediately impacted by the use of coding assistants—such as PR merge rate, review time, and task throughput. 

Best practice: Watch PR merge rate closely. This metric measures the throughput of pull requests merged per developer, on average, per month. Expect this metric increase for developers with Copilot. 

Best practice: Prepare reviewers for increased workloads in advance. Many organizations witness a negative increase in PR Review Time. It may be helpful to revisit SLAs to ensure everyone is on the same page, and set reminders for overdue code reviews. Additionally, as collecting qualitative feedback on AI-augmented changes can provide valuable insights, encourage reviewers to share their thoughts and feedback with program champions.

gauge showing GitHub Copilot Before and After Metrics: PR Review Time

Best practice: Look beyond PR metrics. Introduce data from task management tools like Jira, Azure Devops, or Asana to observe any notable differences in throughput and velocity between the two cohorts. 

bar graph showing GitHub Copilot Before and After Metrics: Task Throughput

Best practice: Balance speed and impact on quality. Monitor quality metrics from static code analysis tools, like SonarQube, or security findings from GitHub Advanced Security to monitor PR Test Coverage, Code Smells, and Number of Vulnerabilities for the cohorts. 

Track leading indicators of productivity improvements

By analyzing data from the GitHub Copilot cohort, you can evaluate performance changes they’re experiencing over time. It’s essential to know which KPIs have increased, decreased, or stayed the same. This data can be used as benchmarks for future rollouts. 

Benchmark: Organizations often see a significant decrease in PR size (up to 90%) and an increase in PR merge rate (up to 25%), while code reviews can become a bottleneck, rising by as much as 20%. 

Best practice: Pay extra attention to power users. When comparing before-and-after metrics, take a close look at power users, your heaviest Copilot adopters. Insights from how their productivity is changing can help project what to expect with higher general usage. 

Learning to run: Transforming individual GitHub Copilot advantages into collective impact

By implementing these best practices during the Learn phase, you’ll be capitalizing on the initial advantages gained from GitHub Copilot and amplifying the impact for teams across your organization. 

Though you never really stop learning and iterating, after 3–6 months, you’ll enter the third stage of the Launch-Learn-Run framework. In our next article, we explore the Run stage, where you’ll examine downstream impacts and collective benefits of GitHub Copilot.

Neely Dunlap

Neely Dunlap

Neely is a content marketer and marketing coordinator at Faros AI.

Connect
AI Is Everywhere. Impact Isn’t.
75% of engineers use AI tools—yet most organizations see no measurable performance gains.

Read the report to uncover what’s holding teams back—and how to fix it fast.
AI Productivity Paradox Report 2025
Discover the Engineering Productivity Handbook
How to build a high-impact program that drives real results.

What to measure and why it matters.

And the 5 critical practices that turn data into impact.
The cover of The Engineering Productivity Handbook on a turquoise background
Want to learn more about Faros AI?

Fill out this form and an expert will reach out to schedule time to talk.

Loading calendar...
An illustration of a lighthouse in the sea

Thank you!

A Faros AI expert will reach out to schedule a time to talk.
P.S. If you don't see it within one business day, please check your spam folder.
Oops! Something went wrong while submitting the form.

More articles for you

Editor's Pick
DevProd
Guides
6
MIN READ

Engineering Team Metrics: How Software Engineering Culture Shapes Performance

Discover which engineering team metrics to track based on your software engineering culture. Learn how cultural values determine the right measurements for your team's success.
August 26, 2025
Editor's Pick
DevProd
Guides
10
MIN READ

Choosing the Best Engineering Productivity Metrics for Modern Operating Models

Engineering productivity metrics vary by operating model. Compare metrics for remote, hybrid, outsourced, and distributed software engineering teams.
August 26, 2025
Editor's Pick
DevProd
Guides
10
MIN READ

How to Choose the Right Software Engineering Metrics for Every Company Stage

Discover the best software engineering metrics for startups, scale-ups, and enterprises. Learn how to choose metrics in software engineering by company stage.
August 25, 2025

See what Faros AI can do for you!

Global enterprises trust Faros AI to accelerate their engineering operations. Give us 30 minutes of your time and see it for yourself.

Salespeak