Key Benchmarks and Considerations for High-Performing Developer Teams
August 24, 2022

Michael Stahnke
CircleCI

In software delivery, there is no question that speed is important. When software teams move fast, good things happen and business value is delivered more frequently.

But, speed comes with a tradeoff: complexity.

As this complexity grows, how can engineering teams succeed?

After analyzing millions of workflows from more than 50,000 organizations across the world, I've outlined some ways teams can start optimizing their software delivery for high performance.

Identify and Meet These 4 Benchmarks

To help teams optimize their software operations for efficiency, CircleCI's latest State of Software Delivery Report examined more than two years of data from over a quarter of a billion workflows, representing more than 50,000 organizations, to gain insights into the DevOps practices used by software teams globally. As a result, the research identified four key benchmarks that the most successful engineering teams routinely meet:

Throughput: Prioritize being in a state of deploy-readiness state most or all of the time, rather than the number of workflows run.

Duration: Reach workflow durations between five to ten minutes on average.

Mean Time to Recovery: Recover from any failed runs by fixing or reverting in under an hour.

Success Rate: Achieve success rates above 90% for the default branch of an application.

Every software team is different. However, the software delivery patterns observed on our platform, especially the data points from top delivery teams, show key similarities that suggest valuable benchmarks for teams to use as goals.

Now let's break down what these four benchmarks really mean.

The number of workflow runs matters less than being in a deploy-ready state most, if not all, of the time. Rather than the number of workflow runs, the most successful teams prioritize being deploy-ready.

The second item that teams should focus on is Duration, which is the time it takes for a workflow to run. Most successful teams achieve workflow durations of five to ten minutes on average.

Third, Mean Time to Recovery describes what it takes for a workflow to become successful again after a failure has occurred. The data shows teams that recover from failed runs in under an hour are the most resilient.

And finally, Success Rate, which is the number of successful runs divided by the total number of runs over a period of time. The most successful engineering teams achieve success rates above 90%.

Prioritize Team Structure and Culture

Prioritizing team structure and culture is essential to improving software delivery metrics. While the ideal team structure and culture will vary depending on the organizational goals, keeping developers in flow is essential to keeping them as productive as possible. That means scheduling meetings at times that don't conflict with peak productivity hours, which the data shows is between 6 a.m. and 7 a.m. PR on Wednesdays.

It is equally important to determine the number of people on your team. Three out of four of our key metrics show a correlation between larger team size and better engineering performance. The research shows the ideal number of code contributors to aim for is between five and twenty, depending on your team's goals, the scope of your responsibilities, as well as other variables. A larger team is also the best way to avoid burnout, and during a time when developer talent is coveted is especially important to consider.

Test, Test, Test

Regardless of your team size, teams prioritizing test-driven development (TDD) can confidently rely on their tooling during market swings, seasonal fluctuations, and times of uncertainty — such as the pandemic. TDD helps companies ensure bad code gets resolved and that organizations can remain safe and resilient.

TDD includes extensive testing, quality checks, and systems that prevent bad code from being put into production. For example, if bad code gets written into your pipeline, TDD can act as a fail-safe when headcount is low. It's the key to preventing bad code from being put into production and staying competitive, regardless of team size.

Great software delivery is a constant loop, not a linear process. The goal for developer teams isn't to make updates to your application, but to constantly innovate on your software while preventing the introduction of faulty changes. Great developer teams that meet the benchmarks outlined above are key to helping businesses differentiate from their competitors and deliver digital products to consumers as fast as the market demands and as often as it changes.

Michael Stahnke is VP of Platform at CircleCI
Share this

Industry News

December 19, 2024

Check Point® Software Technologies Ltd. has been recognized as a Leader in the 2024 Gartner® Magic Quadrant™ for Email Security Platforms (ESP).

December 19, 2024

Progress announced its partnership with the American Institute of CPAs (AICPA), the world’s largest member association representing the CPA profession.

December 18, 2024

Kurrent announced $12 million in funding, its rebrand from Event Store and the official launch of Kurrent Enterprise Edition, now commercially available.

December 18, 2024

Blitzy announced the launch of the Blitzy Platform, a category-defining agentic platform that accelerates software development for enterprises by autonomously batch building up to 80% of software applications.

December 17, 2024

Sonata Software launched IntellQA, a Harmoni.AI powered testing automation and acceleration platform designed to transform software delivery for global enterprises.

December 17, 2024

Sonar signed a definitive agreement to acquire Tidelift, a provider of software supply chain security solutions that help organizations manage the risk of open source software.

December 17, 2024

Kindo formally launched its channel partner program.

December 16, 2024

Red Hat announced the latest release of Red Hat Enterprise Linux AI (RHEL AI), Red Hat’s foundation model platform for more seamlessly developing, testing and running generative artificial intelligence (gen AI) models for enterprise applications.

December 16, 2024

Fastly announced the general availability of Fastly AI Accelerator.

December 12, 2024

Amazon Web Services (AWS) announced the launch and general availability of Amazon Q Developer plugins for Datadog and Wiz in the AWS Management Console.

December 12, 2024

vFunction released new capabilities that solve a major microservices headache for development teams – keeping documentation current as systems evolve – and make it simpler to manage and remediate tech debt.

December 11, 2024

CyberArk announced the launch of FuzzyAI, an open-source framework that helps organizations identify and address AI model vulnerabilities, like guardrail bypassing and harmful output generation, in cloud-hosted and in-house AI models.

December 11, 2024

Grid Dynamics announced the launch of its developer portal.

December 10, 2024

LTIMindtree announced a strategic partnership with GitHub.