17 DevOps Metrics You Should Be Tracking

The metrics we choose can expose problems — or hide them behind irrelevant data. Here are 17 metrics that will help you grow.

Productivity in software development has always been tricky to measure. Unlike in other industries, the act of programming is not something that’s easy to parallelize. The development process is unique in that it requires a diverse mix of technical and communications skills, which calls for a set of specialized metrics to keep track of the team’s vitals.

The pulse of software development

Not all metrics were created equal. Depending on the context, some are more useful than others. The things we choose to measure can help us find problems or obscure them behind irrelevant data and non-productive goals.

When it comes to deciding which metrics to keep track of, we should consider a few points:

  1. People don't act the same when they feel observed. This is called the Hawthorne effect, and it can create undue pressure. It's best to keep the metrics non-personal and anonymous where possible.
  2. The first point also means that metrics should only be used to track a team's progress over time, and not to compare teams or individuals.
  3. Putting too much emphasis on hitting an arbitrary number creates incentives to game the system. Dave Farley and Jez Humble had this to say on that subject:

"Measure the lines of code, and developers will write many short lines of code. Measure the number of defects fixed, and testers will log bugs that could be fixed by a quick discussion with a developer.”

Continuous Delivery: Reliable Software Releases through Build, Test, and Deployment Automation

So, before choosing the metrics you want to use to follow your team's progress, everyone should know that their only purpose is to track progress and identify problems. They are not meant to commend or chastise individuals.

A metric dashboard showing the project score and health.
A dashboard with all the chosen metrics should be created, and it should be visible to everyone on the team.

Four DORA metrics

DORA metrics are the principal tool that we have to measure software development. They consist of four benchmarks:

  • Deployment frequency (DF): how often an organization successfully releases a product to the users or deploys it to production.
  • Lead time to changes (LT): the amount of time it takes a commit to reach production or release.
  • Mean time to restore service (MTTR): how long it takes an organization to recover from a failure in production.
  • Change failure rate (CFR): the percentage of releases or deployments that cause a failure in production.

Development teams can be ranked on one of four levels: Low, Medium, High, and Elite.

MetricLowMediumHighElite
DFfewer than 1 per 6 months1 per month to 1 per 6 months1 per week to 1 per monthOn demand (multiple deploys per day)
LTmore than 6 months1 month to 6 months1 day to 1 weekLess than 1 hour
MTTRmore than 6 months1 day to 1 weekLess than a dayLess than 1 hour
CFR16 to 30%16 to 30%16 to 30%0 to 15%

Year after year, the DORA research team has proven that a high DORA score is a predictable indicator of high performance. As a result, they should be included in any measurement strategy involving software development.

Cycle time

Along with DORA, cycle time is another principal indicator of productivity. It is defined as the average time between the moment we decide to add a feature and its deployment or release to the public or customer.

The stages of development. From left to right: feature approved, feature design, coding, deployment. The cycle time spans all these. While the lead time to changes spans from coding to deployment.
Cycle time spans the entirety of feature development; from inception to reality. Lead time to changes begins ticking when the first line of code for a feature is committed.

A fast cycle time means a team can consistently deliver features at a sustained rate.

Quality

Quality implies different things to different people. While some teams emphasize adhering to style rules, others might be more concerned with security risks or maintaining an enjoyable user experience. What matters is that the team is in agreement regarding what quality entails for them.

We can use a mix of parameters to estimate the quality of the code. That which does not meet a predetermined quality bar should cause the CI pipeline to fail. Some valuable indicators are:

  • Number of vulnerabilities.
  • Violation of style guidelines.
  • Code coverage.
  • Number of stale branches.
  • Cyclomatic complexity.
  • Broken architectural constraints. For instance, making sure that code in one module does not reference classes in another module.

Customer feedback

Customer feedback can come in many forms, such as tickets opened, usage patterns, mentions on social media, and information gleaned from Net Promoter Score (NPS) surveys. The specifics vary depending on the business and product, but we must have the voice of the customer represented in some concrete form because, at the end of the day, they pay the bills.

A short NPS survey asking for feedback.
Are users happy with the product?

Employee satisfaction

Our users and customers are not the only ones whose well-being we must tend to. Developers, testers, quality and business analysts, product managers, and managers are crucial as well, because we need them all to make a great product. The best ideas come from optimistic, confident, and well-rested minds.

Employee satisfaction is affected by various factors, which we should measure in some way:

  • How comprehensive and updated is the documentation?
  • How easy is it to onboard a new developer?
  • Do employees feel their voices are heard?
  • How is the work/life balance? Is anyone burning out?
  • Is the workplace a safe environment to take chances and experiment?
  • Do employees have the right tools to do their jobs?
  • Do they feel they can offer constructive criticism safely?

Average CI duration

Software development is an exercise in experimentation — we make small changes and see how they work out. The feedback from the CI pipeline ultimately determines if a change stays in the codebase.

Working in small increments becomes painful when the CI/CD process is slow, because developers must either wait to see the results or move on and try to remember to return to the pipeline when the results are in. In either case, it is very difficult to keep up the creative flow.

CI duration is equal to total CI runtime divided by total CI runs.
CI duration formula

The CI pipeline's average duration should be measured in minutes. We should aim for less than 10 minutes in order to keep developers engaged and code flowing. If your pipelines take too long, check out Semaphore’s test optimization guide.

CI runs per day

This is the number of CI pipeline executions per day. We want to keep this figure high — at least 4 or 5 runs per active developer — because it implies that developers trust and depend on the CI/CD process.

When the CI runs per day decreases, it might be caused by a slow or awkward-to-use CI/CD system.

CI mean time to recovery (MTTR)

We cannot test, release, or deploy when the build is not working. In such cases, everyone should stop what they are doing and focus on restoring the build. Mean time to recovery measures how long, on average, it takes a team to fix a broken CI build. We're typically only concerned with the main branch when measuring this metric.

Long recovery times signal that we need to work on making the CI/CD process more robust. We must also ensure that the habit of prioritizing the fix of the CI build is ingrained in the team’s culture.

CI test failure rate

Measures how often the CI pipeline fails due to a failed test. Tests are a safety net, so there’s nothing wrong with them failing. Nonetheless, developers should run the tests on their machines before committing code. If the failure rate is too high, it might indicate that developers are finding it hard to run tests locally.

CI test failure rate is equal to test failed divided by total tests executed.
CI test failure rate formula

CI success rate

The CI success rate is the number of successful CI runs divided by the total number of runs. A low success rate indicates that the CI/CD process is brittle, needs more maintenance, or that developers are merging untested code too often.

CI success rate is equal to successful CI run divided by all runs.
CI success rate

Flakiness

Flakiness indicates how brittle the CI pipeline is. A flaky build fails or succeeds randomly for no apparent reason. Flakiness is caused by flaky tests or an unreliable CI/CD platform. Flaky tests negatively impact CI run time, success rate, and time to recovery.

The Semaphore test summary screen.
The Test Summary tab shows flaky and slow tests.

Coverage

Code coverage is the percentage of code that is covered by the test suite. This is a bit controversial since it is a metric that has been known to be frequently misused. For example, requiring 100% coverage does not raise quality — on the contrary, it leads to unnecessary testing of trivial code.

Like anything else, coverage is useful when used in moderation. For instance, a project with 5% coverage is undoubtedly under tested to the point that the outcomes of the tests aren’t showing us much.

Defect Escape Ratio

Measures the number of errors that were not detected by the CI/CD process. A high value means that testing is inadequate. In this case, we should check the coverage value and then re-evaluate how the test suite is structured. We might need more kinds of tests in our test suite.

Defect Escape Ratio is equal to errors found in production divided by all errors found.
Defect Escape Ratio

Uptime

Uptime is the percentage of time the application is available. The higher it is, the less outage there was in a specific period. For example, a 99.9% uptime amounts to 8 hours and 45 minutes of downtime per year. This operational metric should always be part of our measurements as we risk losing customers every time the site or application is down.

A low uptime value points to problems in the infrastructure, the code, and/or the deployment process.

UptimeTotal Yearly Downtime
99.9 %8h 45m 56s
99.99 %52m 35s
99.999 %5m 15s
99.9999 %31s

Service level indicator

Businesses signing service level agreements (SLAs) must pay attention to uptime in order to avoid fines or other penalties. The service level indicator (SLI) contrasts actual application performance or uptime with predetermined standards.

Even when SLAs are not in effect, a company can establish an internal service level objective (SLO), which accomplishes the same function.

A graph showing the SLI as time goes by. The line crosses below the SLA line at some point, triggering some penalties.
SLI shows reality versus SLA or SLO.

Mean time to detection

This is the average time a problem persisted in production before it was detected and assigned to the appropriate team. We can measure it as the time since the problem began until an issue or ticket was raised. Mean time to detection directly correlates to how comprehensive monitoring is and how effective notifications are.

Mean time between failures

Measures how often a system or subsystem fails on average. It’s a metric suitable for measuring the stability of the application’s subcomponents. It can help us determine which parts require refactoring.

The application is modified as time goes by. On the left, the application is missing a feature. Then, the feature is added and works. Finally, the feature breaks. The time between feature is added, and it breaks is the time between failures.
Time between failures.

Metrics just measure symptoms

Metrics are the vital signs of your project. A poor metric is a symptom, not the disease. They point out the presence of a problem but do not say anything definite about the underlying cause. While it might be tempting to fix a problem by “managing” the variables underneath a metric, doing so is akin to self-medicating — it only succeeds in hiding the symptom. Like any good doctor, a good engineer investigates, and proposes solutions and confirms their effectiveness by checking to see if the metric has improved.

Thanks for reading, and happy measuring!