Key Benchmarks and Considerations for High-Performing Developer Teams
August 24, 2022

Michael Stahnke
CircleCI

In software delivery, there is no question that speed is important. When software teams move fast, good things happen and business value is delivered more frequently.

But, speed comes with a tradeoff: complexity.

As this complexity grows, how can engineering teams succeed?

After analyzing millions of workflows from more than 50,000 organizations across the world, I've outlined some ways teams can start optimizing their software delivery for high performance.

Identify and Meet These 4 Benchmarks

To help teams optimize their software operations for efficiency, CircleCI's latest State of Software Delivery Report examined more than two years of data from over a quarter of a billion workflows, representing more than 50,000 organizations, to gain insights into the DevOps practices used by software teams globally. As a result, the research identified four key benchmarks that the most successful engineering teams routinely meet:

Throughput: Prioritize being in a state of deploy-readiness state most or all of the time, rather than the number of workflows run.

Duration: Reach workflow durations between five to ten minutes on average.

Mean Time to Recovery: Recover from any failed runs by fixing or reverting in under an hour.

Success Rate: Achieve success rates above 90% for the default branch of an application.

Every software team is different. However, the software delivery patterns observed on our platform, especially the data points from top delivery teams, show key similarities that suggest valuable benchmarks for teams to use as goals.

Now let's break down what these four benchmarks really mean.

The number of workflow runs matters less than being in a deploy-ready state most, if not all, of the time. Rather than the number of workflow runs, the most successful teams prioritize being deploy-ready.

The second item that teams should focus on is Duration, which is the time it takes for a workflow to run. Most successful teams achieve workflow durations of five to ten minutes on average.

Third, Mean Time to Recovery describes what it takes for a workflow to become successful again after a failure has occurred. The data shows teams that recover from failed runs in under an hour are the most resilient.

And finally, Success Rate, which is the number of successful runs divided by the total number of runs over a period of time. The most successful engineering teams achieve success rates above 90%.

Prioritize Team Structure and Culture

Prioritizing team structure and culture is essential to improving software delivery metrics. While the ideal team structure and culture will vary depending on the organizational goals, keeping developers in flow is essential to keeping them as productive as possible. That means scheduling meetings at times that don't conflict with peak productivity hours, which the data shows is between 6 a.m. and 7 a.m. PR on Wednesdays.

It is equally important to determine the number of people on your team. Three out of four of our key metrics show a correlation between larger team size and better engineering performance. The research shows the ideal number of code contributors to aim for is between five and twenty, depending on your team's goals, the scope of your responsibilities, as well as other variables. A larger team is also the best way to avoid burnout, and during a time when developer talent is coveted is especially important to consider.

Test, Test, Test

Regardless of your team size, teams prioritizing test-driven development (TDD) can confidently rely on their tooling during market swings, seasonal fluctuations, and times of uncertainty — such as the pandemic. TDD helps companies ensure bad code gets resolved and that organizations can remain safe and resilient.

TDD includes extensive testing, quality checks, and systems that prevent bad code from being put into production. For example, if bad code gets written into your pipeline, TDD can act as a fail-safe when headcount is low. It's the key to preventing bad code from being put into production and staying competitive, regardless of team size.

Great software delivery is a constant loop, not a linear process. The goal for developer teams isn't to make updates to your application, but to constantly innovate on your software while preventing the introduction of faulty changes. Great developer teams that meet the benchmarks outlined above are key to helping businesses differentiate from their competitors and deliver digital products to consumers as fast as the market demands and as often as it changes.

Michael Stahnke is VP of Platform at CircleCI
Share this

Industry News

April 03, 2025

StackGen has partnered with Google Cloud Platform (GCP) to bring its platform to the Google Cloud Marketplace.

April 03, 2025

Tricentis announced its spring release of new cloud capabilities for the company’s AI-powered, model-based test automation solution, Tricentis Tosca.

April 03, 2025

Lucid Software has acquired airfocus, an AI-powered product management and roadmapping platform designed to help teams prioritize and build the right products faster.

April 03, 2025

AutonomyAI announced its launch from stealth with $4 million in pre-seed funding.

April 02, 2025

Kong announced the launch of the latest version of Kong AI Gateway, which introduces new features to provide the AI security and governance guardrails needed to make GenAI and Agentic AI production-ready.

April 02, 2025

Traefik Labs announced significant enhancements to its AI Gateway platform along with new developer tools designed to streamline enterprise AI adoption and API development.

April 02, 2025

Zencoder released its next-generation AI coding and unit testing agents, designed to accelerate software development for professional engineers.

April 02, 2025

Windsurf (formerly Codeium) and Netlify announced a new technology partnership that brings seamless, one-click deployment directly into the developer's integrated development environment (IDE.)

April 02, 2025

Opsera raised $20M in Series B funding.

April 02, 2025

The Cloud Native Computing Foundation® (CNCF®), which builds sustainable ecosystems for cloud native software, is making significant updates to its certification offerings.

April 01, 2025

The Cloud Native Computing Foundation® (CNCF®), which builds sustainable ecosystems for cloud native software, announced the Golden Kubestronaut program, a distinguished recognition for professionals who have demonstrated the highest level of expertise in Kubernetes, cloud native technologies, and Linux administration.

April 01, 2025

Red Hat announced new capabilities and enhancements for Red Hat Developer Hub, Red Hat’s enterprise-grade internal developer portal based on the Backstage project.

April 01, 2025

Platform9 announced that Private Cloud Director Community Edition is generally available.

March 31, 2025

Sonatype expanded support for software development in Rust via the Cargo registry to the entire Sonatype product suite.

March 31, 2025

CloudBolt Software announced its acquisition of StormForge, a provider of machine learning-powered Kubernetes resource optimization.