Key Benchmarks and Considerations for High-Performing Developer Teams
August 24, 2022

Michael Stahnke
CircleCI

In software delivery, there is no question that speed is important. When software teams move fast, good things happen and business value is delivered more frequently.

But, speed comes with a tradeoff: complexity.

As this complexity grows, how can engineering teams succeed?

After analyzing millions of workflows from more than 50,000 organizations across the world, I've outlined some ways teams can start optimizing their software delivery for high performance.

Identify and Meet These 4 Benchmarks

To help teams optimize their software operations for efficiency, CircleCI's latest State of Software Delivery Report examined more than two years of data from over a quarter of a billion workflows, representing more than 50,000 organizations, to gain insights into the DevOps practices used by software teams globally. As a result, the research identified four key benchmarks that the most successful engineering teams routinely meet:

Throughput: Prioritize being in a state of deploy-readiness state most or all of the time, rather than the number of workflows run.

Duration: Reach workflow durations between five to ten minutes on average.

Mean Time to Recovery: Recover from any failed runs by fixing or reverting in under an hour.

Success Rate: Achieve success rates above 90% for the default branch of an application.

Every software team is different. However, the software delivery patterns observed on our platform, especially the data points from top delivery teams, show key similarities that suggest valuable benchmarks for teams to use as goals.

Now let's break down what these four benchmarks really mean.

The number of workflow runs matters less than being in a deploy-ready state most, if not all, of the time. Rather than the number of workflow runs, the most successful teams prioritize being deploy-ready.

The second item that teams should focus on is Duration, which is the time it takes for a workflow to run. Most successful teams achieve workflow durations of five to ten minutes on average.

Third, Mean Time to Recovery describes what it takes for a workflow to become successful again after a failure has occurred. The data shows teams that recover from failed runs in under an hour are the most resilient.

And finally, Success Rate, which is the number of successful runs divided by the total number of runs over a period of time. The most successful engineering teams achieve success rates above 90%.

Prioritize Team Structure and Culture

Prioritizing team structure and culture is essential to improving software delivery metrics. While the ideal team structure and culture will vary depending on the organizational goals, keeping developers in flow is essential to keeping them as productive as possible. That means scheduling meetings at times that don't conflict with peak productivity hours, which the data shows is between 6 a.m. and 7 a.m. PR on Wednesdays.

It is equally important to determine the number of people on your team. Three out of four of our key metrics show a correlation between larger team size and better engineering performance. The research shows the ideal number of code contributors to aim for is between five and twenty, depending on your team's goals, the scope of your responsibilities, as well as other variables. A larger team is also the best way to avoid burnout, and during a time when developer talent is coveted is especially important to consider.

Test, Test, Test

Regardless of your team size, teams prioritizing test-driven development (TDD) can confidently rely on their tooling during market swings, seasonal fluctuations, and times of uncertainty — such as the pandemic. TDD helps companies ensure bad code gets resolved and that organizations can remain safe and resilient.

TDD includes extensive testing, quality checks, and systems that prevent bad code from being put into production. For example, if bad code gets written into your pipeline, TDD can act as a fail-safe when headcount is low. It's the key to preventing bad code from being put into production and staying competitive, regardless of team size.

Great software delivery is a constant loop, not a linear process. The goal for developer teams isn't to make updates to your application, but to constantly innovate on your software while preventing the introduction of faulty changes. Great developer teams that meet the benchmarks outlined above are key to helping businesses differentiate from their competitors and deliver digital products to consumers as fast as the market demands and as often as it changes.

Michael Stahnke is VP of Platform at CircleCI
Share this

Industry News

May 09, 2024

Red Hat and Elastic announced an expanded collaboration to deliver next-generation search experiences supporting retrieval augmented generation (RAG) patterns using Elasticsearch as a preferred vector database solution integrated on Red Hat OpenShift AI.

May 09, 2024

Traceable AI announced an Early Access Program for its new Generative AI API Security capabilities.

May 09, 2024

StackHawk announced a new integration with Microsoft Defender for Cloud to help organizations build software more securely.

May 08, 2024

MacStadium announced that it has obtained Cloud Security Alliance (CSA) Security, Trust & Assurance Registry (STAR) Level 1, meaning that MacStadium has publicly documented its compliance with CSA’s Cloud Controls Matrix (CCM), and that it joined the Cloud Security Alliance (CSA), the world’s leading organization dedicated to defining and raising awareness of best practices to help ensure a secure cloud computing environment.

May 08, 2024

The Cloud Native Computing Foundation® (CNCF®) released the two-day schedule for CloudNativeSecurityCon North America 2024 happening in Seattle, Washington from June 26-27, 2024.

May 08, 2024

Sumo Logic announced new AI and security analytics capabilities that allow security and development teams to align around a single source of truth and collect and act on data insights more quickly.

May 08, 2024

Red Hat is announcing an optional additional 12-month EUS term for OpenShift 4.14 and subsequent even-numbered Red Hat OpenShift releases in the 4.x series.

May 08, 2024

HAProxy Technologies announced the launch of HAProxy Enterprise 2.9.

May 08, 2024

ArmorCode announced the general availability of AI Correlation in the ArmorCode ASPM Platform.

May 08, 2024

Octopus Deploy launched new features to help simplify Kubernetes CD at scale for enterprises.

May 08, 2024

Cequence announced multiple ML-powered advancements to its Unified API Protection (UAP) platform.

May 07, 2024

Oracle announced plans for Oracle Code Assist, an AI code companion, to help developers boost velocity and enhance code consistency.

May 07, 2024

New Relic launched Secure Developer Alliance.

May 07, 2024

Dynatrace is enhancing its platform with new Kubernetes Security Posture Management (KSPM) capabilities for observability-driven security, configuration, and compliance monitoring.

May 07, 2024

Red Hat announced advances in Red Hat OpenShift AI, an open hybrid artificial intelligence (AI) and machine learning (ML) platform built on Red Hat OpenShift that enables enterprises to create and deliver AI-enabled applications at scale across hybrid clouds.