Beyond the Hype: New Research Exposes the Messy Reality of Software Cycle Time Metrics
Share this article
For years, engineering leaders have treated cycle time—the duration from ticket creation to resolution—as a holy grail of developer productivity measurement. New research published in Empirical Software Engineering dismantles this oversimplified view through rigorous analysis of 55,619 development tickets from 11,398 contributors across 216 organizations. The findings reveal a metric drowning in noise, where even statistically significant factors explain only fragments of the complex reality of software delivery.
The Illusion of Control
The Bayesian hierarchical modeling study confirmed expected relationships:
- Coding days: Developers with 1+ additional coding day/week saw ~2-day shorter cycle times (median 13 days)
- Pull requests: More merged PRs correlated with faster ticket resolution
- Collaboration: Higher network centrality (shared PR contributions) reduced cycle time
Yet these effects pale against the staggering variability observed. Between-person differences accounted for 65% of cycle time variance (SD=0.645), while organizational differences explained 47% (SD=0.474). As the researchers note: "Any single observation of cycle time is a very noisy indication of what is typical."
Organization sizes in the study (median=130 developers) showing diverse participation across industries
The Perverse Metrics Paradox
Contradictory findings emerged that defy simplistic interpretations:
- Defect tickets: Within-person increases in bug work correlated with shorter cycle times, while developers consistently handling more defects had longer cycles
- PR comments: More comments per pull request consistently predicted longer cycle times—potentially indicating complex/problematic changes
- Quarterly deadlines: Cycle times slightly decreased near quarter-ends, suggesting deadline-driven prioritization
Lead researcher Dr. John Flournoy warns: "Resolution increases with more measurements across time and people. We do not know of a measure with properties allowing reliable individual assessment in short timeframes."
Why the "10x Engineer" Myth Persists
The data exposes fundamental flaws in individual performance comparisons:
1. Task inequality: Tickets vary wildly in complexity (found: 31x difference in cycle time variance)
2. Context collapse: Organizational processes explain nearly half of cycle time differences
3. Measurement artifacts: 17% of monthly median cycle times fell outside 95% prediction intervals
Effect size heterogeneity across organizations—identical changes yield different impacts
A Path Forward
The study advocates systemic interventions:
- Prioritize psychological affordances: Developers reporting higher autonomy show 23% less cycle time volatility
- Embrace team-level analysis: 72% of variability resides above individual level
- Combine metrics: Pair cycle time with developer self-reports on blockers
As the paper concludes: "Improving software delivery velocity requires systems-level thinking rather than individual-focused interventions. There are no silver bullets—only context-aware tradeoffs."
Source: Flournoy, J.C. et al. (2025). "No Silver Bullets: Why Understanding Software Cycle Time Is Messy, Not Magic." Empirical Software Engineering 30(174). DOI:10.1007/s10664-025-10735-w