How Developers Game Sprint Metrics (and How to Prevent It)
Engineering managers rely heavily on sprint metrics such as story points, velocity, and burndown charts to track progress, predict delivery, and improve team performance. While these metrics can provide powerful insights into a team's effectiveness, relying too heavily on them can inadvertently encourage developers to manipulate data rather than genuinely improve productivity or quality. This practice, known as gaming, undermines trust and weakens the value of agile metrics.
To foster genuine improvement, engineering leaders must recognize how and why developers game metrics, how to spot it early, and how to prevent it altogether.
Why Developers Game Metrics
Developers typically game sprint metrics when they feel pressure to meet unrealistic targets or when metrics are used to evaluate their individual performance or job security. Common motivations include:
- Performance Anxiety: Developers fear negative evaluations or career consequences if they fail to hit specific numeric targets such as velocity or completed story points.
- Misaligned Incentives: If metrics are tied directly to bonuses or promotions, developers may focus more on improving their numbers rather than delivering real customer value.
- Poorly Defined Metrics: When teams rely on metrics that don't accurately represent productivity or quality, developers often try to manipulate these numbers to appear successful on paper.
Martin Fowler describes this behavior clearly, noting that when metrics become targets, teams inevitably alter their behavior to improve those numbers, even at the expense of genuine productivity or quality.
How Developers Game Sprint Metrics
Below are some common ways developers manipulate sprint metrics:
Inflating Story Points
Teams under pressure to show increased velocity may inflate story point estimates, assigning higher points to tasks than warranted. Over time, story point values creep upward, artificially boosting velocity without genuine improvement.
Over-Splitting User Stories
Developers might split stories excessively into tiny, trivial tasks to increase completed task counts. The result appears as improved throughput, despite no real increase in actual delivered value or efficiency.
Marking Unfinished Work as Done
To avoid rollover tickets, developers sometimes mark incomplete tasks as complete, intending to finish them later quietly. While this temporarily improves sprint completion rates, it creates hidden debt and inaccurate metrics about actual progress.
Favoring Easy or Low-Value Tasks
Under pressure to deliver high numbers, teams may prioritize simple, easy-to-complete tasks over more important but challenging work. This practice inflates productivity numbers without increasing customer value.
Relaxing the Definition of Done
Teams may loosen their criteria for marking stories as complete, such as reducing testing standards or skipping documentation steps. While it helps show tasks as done faster, it often results in lower-quality software and technical debt.
Changing How Metrics Are Calculated
Teams occasionally adjust how they measure velocity or task completion, such as excluding certain categories of work or extending sprint lengths. This type of metric manipulation creates the appearance of successful sprints without actual performance improvements.
Google’s research on code quality and testing emphasizes that reducing testing rigor or skipping code reviews to boost metrics typically harms software quality over the long run.
How to Identify Metric Gaming
Managers can identify potential gaming by observing common patterns:
Unexpected Spikes in Sprint Velocity
Sudden jumps in velocity or consistently rising velocity numbers without clear process improvements suggest possible estimate inflation or metric manipulation.
Frequent Last-Minute Scope Adjustments
If stories regularly shift in or out of sprints at the last minute, or if incomplete stories suddenly appear as done right before sprint end, teams may be artificially inflating completion rates.
Increasing Story Point Values Over Time
Reviewing historical sprint data to identify a steady increase in average story point size could reveal systematic inflation aimed at boosting velocity.
High Counts of Trivial Tasks
Regularly including trivial tasks such as minor refactoring or administrative work with story points suggests efforts to pad sprint metrics.
Team Jokes or Comments About Gaming Metrics
Developers making cynical comments or openly discussing "hitting numbers" rather than delivering value is often a sign metrics are being manipulated.
Preventing Sprint Metric Gaming
Engineering leaders can reduce or prevent metric gaming by adopting the following strategies:
Managers should always have the final judgment
Metrics should enhance a manager’s understanding of employee performance, not replace it. Managers should holistically assess employee performance and always have the final say in assessments. This decreases incentives to manipulate metrics.
Do not use sprint metrics as OKRs or performance KPIs
If your organization uses OKRs or another metrics framework that ties KPIs to employee performance, avoid using story points or other sprint metrics at all costs. Instead, focus on metrics that reflect value delivery, like customer satisfaction, new bug rates, etc.
Emphasize Outcomes Rather than Numbers
Clearly communicate that the primary goal is customer value, software quality, and team improvement rather than numeric targets. Remind teams that metrics are tools, not targets.
Avoid Ranking Teams or Individuals Based on Metrics
Comparing developers or teams using metrics encourages unhealthy competition and gaming behaviors. Use metrics as diagnostic tools to identify improvement opportunities, not as performance rankings.
Adopt Multiple, Complementary Metrics
Balance output metrics like velocity with outcome and quality metrics such as customer satisfaction, cycle time, or defect rates. A balanced approach makes it harder for teams to manipulate metrics without being noticed.
Foster Transparency and Open Communication
Create a safe, open culture where teams can honestly discuss missed targets or unexpected outcomes without fear of blame. Transparency reduces pressure and the perceived need to manipulate data.
Regularly Calibrate Estimates and Metrics
Periodically review and recalibrate story point estimates and metric definitions. Encourage estimation workshops to reset baseline assumptions, preventing gradual drift in how metrics are measured.
The DORA research emphasizes the importance of combining several metrics, such as deployment frequency, lead time, and change failure rate, to provide balanced feedback and reduce the temptation to game a single measurement.
Conclusion
Sprint metrics like velocity, story points, and burndown charts can be incredibly useful for agile teams, when used responsibly. Engineering leaders must be aware that when metrics become targets or performance measures, developers may start gaming them to appear successful. By maintaining transparency, emphasizing value over numeric targets, and using multiple balanced metrics, leaders can help teams remain focused on genuine productivity and improvement rather than on manipulating numbers.
Keeping metrics informative rather than punitive fosters a healthy team environment where engineers naturally focus on delivering quality software and customer value. This approach not only provides more accurate sprint data but also creates a stronger, more collaborative engineering culture.
By following these principles, engineering teams and leaders can harness metrics as effective tools for improvement and avoid the pitfalls of metric gaming altogether.