Here's something that surprises many Scrum practitioners: the word "velocity" does not appear anywhere in the Scrum Guide. Not once. It never has.

And yet velocity has become the de facto metric that organisations use to measure team performance, compare teams against each other, set contractual targets, and make staffing decisions. This isn't a minor misunderstanding — it's a measurement dysfunction that distorts incentives, inflates estimates, and disconnects delivery teams from the outcomes that actually matter.

As Scrum.org states plainly: "Velocity is a planning tool rather than a specific measurement. It's an estimate of relative capacity that tends to change over time, and these changes don't necessarily indicate a shift in productivity."

0
The number of times "velocity" appears in the Scrum Guide. It is not a prescribed Scrum metric — it's an emergent practice that has been widely adopted and widely misused.
Source: Scrum Guide (2020). Verified independently. Also noted by Concepts and Beyond, Scrum.org, and others.

Why Velocity Fails as a Performance Metric

Velocity measures the average number of story points completed by a team in a sprint. On the surface, it seems useful. But it has fundamental limitations that make it actively harmful when used beyond its intended purpose of sprint capacity forecasting.

Story points are relative, not absolute. Every team defines their own story point system. There is no standard formula. A team reporting velocity of 34 and a team reporting 13 could be doing equivalent work — or wildly different work. Cross-team comparison is meaningless.

Velocity targets create perverse incentives. Scrum.org documented a case where a Statement of Work mandated 60 story points per sprint. The team responded by filling the backlog with unrelated work items and inflating estimates to hit the target. The metric was met. Value was not delivered.

It measures output, not outcomes. Twenty well-delivered story points that solve a real customer problem are worth infinitely more than forty rushed points that nobody uses. Velocity cannot distinguish between the two.

AI is breaking it further. Scrum.org's 2026 analysis of velocity in the AI era made this point starkly: when AI agents can generate code for a "3-point story" in seconds, velocity jumps from 50 to 5,000 in a sprint. Has the team delivered 100x more value? No. They've broken the metric. As the analysis concluded: "If we cling to Velocity, we are lying to ourselves about our productivity."

The 5 Metrics That Actually Matter

Based on a decade of DORA research (39,000+ professionals), Scrum.org's Evidence-Based Management framework (updated May 2024), and empirical evidence from the broader agile research community, here are five metrics that predict delivery effectiveness far better than velocity.

1

Deployment Frequency

How often your team deploys to production or releases to end users. DORA's research programme has consistently found this to be the single strongest differentiator between high and low-performing organisations.

According to DORA's 2024 report, elite-performing teams deploy on demand — multiple times per day. Low performers deploy between once per month and once every six months. The gap between these groups has widened since 2023, with the low-performing cluster growing from 17% to 25% of respondents.

Why it matters: Frequent deployment forces small batch sizes, which reduces risk, accelerates feedback, and makes integration problems visible early. It's a forcing function for good engineering practices — not just a measurement of speed.
2

Cycle Time (Lead Time for Changes)

The elapsed time from when work begins on an item to when it reaches the customer. DORA defines this specifically as the time from code commit to production deployment. In a broader agile context, it includes the full journey from backlog selection to customer delivery.

This is a fundamentally different measurement from velocity. Velocity tells you how many points you completed. Cycle time tells you how long each piece of work actually took to flow through your system. It exposes queues, handoffs, review bottlenecks, and waiting time that velocity completely hides.

Why it matters: Cycle time is a leading indicator. When it increases, your pipeline is developing friction — even if velocity looks stable. Track it at the item level, not the sprint level, to see real flow patterns.
3

Sprint Goal Achievement Rate

The percentage of sprints where the team achieves its Sprint Goal — not "completed all stories" but achieved the coherent objective that gives the sprint its purpose. The Sprint Goal is explicitly called out in the 2020 Scrum Guide as a commitment for the Sprint Backlog.

This metric reframes success from "did we complete X story points?" to "did we achieve the outcome we committed to?" A team that completes 90% of stories but misses the Sprint Goal has not had a successful sprint. A team that pivots mid-sprint but nails the goal has.

Why it matters: Sprint Goal achievement connects delivery to purpose. It also surfaces whether teams are crafting meaningful Sprint Goals at all — many aren't, which is itself diagnostic information.
4

Change Failure Rate

The percentage of deployments that result in degraded service or require remediation (rollback, hotfix, patch). This is one of DORA's four key metrics and part of their stability measurement.

DORA's 2024 report noted something interesting: for the first time, the medium-performing cluster actually had a lower change failure rate than the high-performing cluster. This decoupling of throughput and stability metrics was unusual in DORA's decade of research. DORA chose to rank high performers based on more frequent deployments even with higher failure rates, but acknowledged the tension.

Why it matters: Change failure rate directly measures quality in production. Combined with deployment frequency, it tells you whether your team is delivering sustainably or accumulating hidden risk. Atlassian recommends tracking DORA metrics together, noting that "a consistently high deployment frequency doesn't tell the whole story if the change failure rate is also consistently high."
5

Unrealised Value (Customer Satisfaction Gap)

This comes directly from Scrum.org's Evidence-Based Management (EBM) framework. Unrealised Value measures the gap between what customers desire and what they currently experience. It's the most strategically important metric on this list — and the one fewest teams track.

The 2024 EBM Guide classifies measures into five types: Input, Activity, Output, Outcome, and Impact. Most teams measure inputs and activities. Unrealised Value operates at the outcome and impact level. It asks: how much potential value exists that we haven't yet captured?

Practical measures include Net Promoter Score trends, feature request analysis, customer churn rates, and competitive gap assessments. The EBM framework positions this alongside Current Value, Time-to-Market, and Ability to Innovate as the four Key Value Areas that give organisations a complete picture of their value delivery.

Why it matters: A team with perfect velocity and zero customer satisfaction improvement is running efficiently in the wrong direction. Unrealised Value is how you know whether you're solving problems that matter.

What About Action Item Completion?

There's a meta-metric that Scrum Masters rarely track but should: the completion rate of retrospective action items. Data from EasyRetro, one of the largest retrospective platforms, shows the average completion rate of retrospective action items across their platform is approximately 0.33%. That's not a typo — less than one percent of improvement actions identified in retrospectives get marked as complete.

Easy Agile's usage data from their TeamRhythm tool tells a more nuanced but still sobering story: teams were completing only 40-50% of their retrospective action items. After they released features specifically designed to surface and track incomplete actions, completion jumped to 65%. The tool didn't change the teams' intentions — it changed their visibility and accountability.

If your retrospective produces three action items every sprint and fewer than half are completed before the next one, you have a continuous improvement system that isn't actually improving anything. Track this.

Key Takeaway

Velocity is a capacity forecasting tool — nothing more. For understanding delivery effectiveness, measure deployment frequency, cycle time, Sprint Goal achievement, change failure rate, and unrealised value. Together, these five metrics connect team-level execution to customer outcomes and strategic goals.

SprintINSite provides automated velocity and throughput tracking, sprint-over-sprint trend analysis, and capacity forecasting inside Jira Cloud — the operational foundation you need before layering on outcome metrics. Try it on the Atlassian Marketplace.

Getting Started: A Four-Week Adoption Plan

Week 1: Start measuring cycle time at the issue level in Jira. Most teams can extract this from existing data without any new tooling. Track it per story, not per sprint.

Week 2: Introduce Sprint Goal achievement tracking. At each Sprint Review, ask the binary question: "Did we achieve the Sprint Goal?" Record yes or no. Track it over time. If your team doesn't have a Sprint Goal, that's your first improvement.

Week 3: If your team deploys to production, start capturing deployment frequency and change failure rate. If you use Jira Service Management or a CI/CD pipeline with logging, this data likely already exists — it just isn't surfaced.

Week 4: Have a conversation with your Product Owner about Unrealised Value. What customer problems remain unsolved? What satisfaction gaps exist? You don't need a formal survey — start with support ticket analysis, feature request themes, and churn data. Establish a baseline.

Ongoing: Stop using velocity for anything other than sprint capacity planning. If it's in a management dashboard, remove it. If it's in a contract, flag it. Every time velocity is used as a performance target, it actively degrades the quality of your delivery.

Ready to measure what matters?

SprintINSite gives Jira Cloud teams automated sprint analytics, velocity trends, and capacity forecasting — the data foundation for moving beyond velocity to outcome-driven delivery. Australian-built for Atlassian teams.

Explore SprintINSite

References & Sources

  1. Scrum Guide (2020). Ken Schwaber & Jeff Sutherland. Velocity is not mentioned. scrumguides.org
  2. Scrum.org. "Myth: Velocity is Productivity." scrum.org
  3. Scrum.org. "Scrum Masters, What Are You Measuring in 2025?" Documents velocity target gaming case. scrum.org
  4. Scrum.org. "From Velocity to Agent Efficiency: Evidence-Based Management for the AI Era." February 2026. scrum.org
  5. Concepts and Beyond. "Velocity: The Misused Metric." April 2025. conceptsandbeyond.com
  6. Google Cloud / DORA. "2024 Accelerate State of DevOps Report." 39,000+ professionals over a decade. dora.dev
  7. DX (formerly GetDX). "Highlights from the 2024 DORA State of DevOps Report." Analysis of cluster shifts. getdx.com
  8. Octopus Deploy. "The 2024 DevOps Performance Clusters." octopus.com
  9. Scrum.org. "Evidence-Based Management Guide." Updated May 2024. scrum.org
  10. Atlassian. "DORA Metrics: How to Measure Open DevOps Success." January 2026. atlassian.com
  11. EasyRetro. "Retrospective Statistics." Average action item completion rate: 0.33%. easyretro.io
  12. Easy Agile. "Why Retrospectives Fail: Fixing Action Item Follow-Through." 40-50% completion rate jumping to 65% with tracking features. easyagile.com