What Are DORA Metrics? A Practical Guide for Engineering Teams
DORA metrics explained with real benchmarks, calculation methods, common mistakes, and how they connect to business outcomes. Based on a decade of research from Google's DORA team.
Your VP of Engineering just asked you to "start tracking engineering metrics." No specifics. No context on what decisions these metrics should inform. Just "we need more visibility into how the team is performing."
So you google "engineering metrics" and get hit with a wall of acronyms: DORA, SPACE, DX, cycle time, velocity, throughput, change failure rate. Some articles say to track lines of code (don't). Others recommend 20+ metrics dashboards that nobody will look at after the first week.
This guide cuts through the noise. We'll cover the engineering metrics that actually drive decisions, how to measure them without creating a surveillance culture, and which ones to ignore entirely.
Key Takeaways: The best engineering metrics measure system performance, not individual output. Start with 3-5 metrics tied to specific decisions, not 20 metrics on a dashboard nobody checks.
Engineering metrics are quantitative measures of how a software engineering team builds, ships, and maintains software. They cover the full development lifecycle: from how quickly code moves from idea to production, to how often deployments break, to how efficiently teams collaborate on code reviews.
The purpose of engineering metrics is not to measure individual developer performance. It's to identify systemic bottlenecks, validate process changes, and give engineering leaders data to make better decisions about staffing, tooling, and priorities.
Think of engineering metrics like vital signs for your development organization. A doctor doesn't use blood pressure to judge whether you're a good person. They use it to diagnose problems and track whether treatments are working. Engineering metrics should work the same way.
Instead of listing 30 metrics, we'll organize them by the question they answer. Pick the questions most relevant to your team, then track only the metrics that inform those decisions.
Cycle Time measures the elapsed time from when work starts (first commit or branch creation) to when it's deployed to production. This is the single most important delivery metric because it captures everything: coding speed, review delays, CI/CD pipeline time, and deployment processes.
Deployment Frequency tracks how often your team ships code to production. This is one of the four DORA metrics and is strongly correlated with overall engineering performance.
PR Throughput measures how many pull requests your team merges per week. It's a useful proxy for delivery pace, especially when combined with PR size data.
Change Failure Rate is the percentage of deployments that cause a failure in production (outage, rollback, hotfix, or degraded service). Another core DORA metric, it measures whether you're shipping fast at the expense of stability.
Mean Time to Recovery (MTTR) measures how long it takes to restore service after a production incident. Fast recovery is often more important than preventing every failure.
Bug Escape Rate tracks the number of bugs found in production versus those caught in QA, staging, or code review. A rising escape rate means your quality gates are leaking.
PR Review Time measures the elapsed time from when a pull request is opened to when it receives its first review. This is often the biggest hidden bottleneck in engineering teams.
PR Size (lines changed per PR) is a leading indicator of review quality and merge speed. Research from Google's engineering practices team shows that PRs over 400 lines receive significantly less thorough reviews.
Work Breakdown tracks the percentage split between new features, bug fixes, maintenance, and technical debt. This is crucial for engineering leaders who need to justify "invisible" work to stakeholders.
Not all metrics are useful. Some actively damage team culture and incentivize the wrong behaviors. Here are the metrics you should avoid:
Tracking lines of code rewards verbosity and penalizes elegant solutions. A developer who deletes 500 lines of dead code while simplifying a module has done more valuable work than someone who added 500 lines of boilerplate. LoC is a meaningless metric that has been debunked by every credible engineering research team, including Google's DORA team and Microsoft Research.
Commit frequency is a function of personal workflow, not productivity. Some developers commit every 20 minutes. Others commit once when the feature is complete. Neither approach is inherently better. Tracking commits per day incentivizes meaningless micro-commits.
Publishing leaderboards of story points completed per developer destroys psychological safety and encourages gaming. Developers will inflate estimates, avoid helping teammates, and cherry-pick easy tickets. Velocity is a team metric, never an individual one.
Tracking time spent coding (via editor plugins or activity monitors) measures input, not output. A senior engineer who spends 2 hours designing the right solution and 1 hour coding it produces far more value than someone who spends 8 hours writing code they'll refactor tomorrow.
Rule of thumb: If a metric can be gamed by an individual in a way that doesn't improve the team's actual output, it's not worth tracking.
The biggest risk with engineering metrics isn't choosing the wrong ones. It's implementing them in a way that makes developers feel monitored rather than supported. Here's how to get it right:
Every metric should be reported at the team level first. Individual breakdowns are useful for 1:1 conversations (e.g., "I noticed your PRs are sitting in review longer than usual — are you blocked on something?") but should never be published to the team or leadership without context.
Before adding a metric, ask: "What decision will this help us make?" If you can't answer that question, don't track it. For example: "Cycle time will help us decide whether to invest in CI pipeline speed" is a good reason. "We want a dashboard" is not.
Metrics should be visible to the entire team, not just management. When developers can see their own team's cycle time and PR review latency, they self-correct. Hidden metrics that only managers see create distrust.
A cycle time of 4 days isn't inherently good or bad. But cycle time increasing from 2 days to 4 days over three sprints is a signal worth investigating. Focus on direction and rate of change, not on hitting arbitrary benchmarks.
Teams that track 20 metrics track nothing. Information overload leads to dashboard fatigue and no action. Start with 3-5 metrics aligned to your team's current challenges. You can always add more later.
Different roles need different views of engineering metrics. Here's what matters most at each level:
| Role | Key Metrics | Why |
|---|---|---|
| CTO / VP Eng | Deployment frequency, change failure rate, work breakdown | Strategic decisions on team investment and reliability |
| Engineering Manager | Cycle time, PR review time, sprint velocity, blockers | Process optimization and team health |
| Tech Lead | PR size, review turnaround, code churn, test coverage | Code quality and review process efficiency |
| Product Manager | Feature vs. bug ratio, sprint completion rate, cycle time | Roadmap planning and delivery predictability |
The most widely adopted framework, developed by Google's DevOps Research and Assessment team. It tracks four metrics: deployment frequency, lead time for changes, change failure rate, and mean time to recovery. These four metrics are statistically validated as predictors of organizational performance across 10+ years of research involving over 39,000 professionals. Read our complete DORA metrics guide for a deep dive.
Developed by researchers at Microsoft, GitHub, and the University of Victoria, SPACE argues that developer productivity can't be captured by a single metric. It proposes five dimensions: Satisfaction and well-being, Performance, Activity, Communication and collaboration, and Efficiency and flow. SPACE is useful as a lens for ensuring you're not over-indexing on any single dimension.
A newer framework focused on three core dimensions: feedback loops (how quickly developers get information about their code), cognitive load (how much mental effort is required to complete tasks), and flow state (how often developers can work without interruptions). DevEx is gaining traction because it focuses on the developer's experience rather than just output metrics.
Here's a practical timeline for rolling out engineering metrics:
Start with cycle time, deployment frequency, and change failure rate — the core DORA metrics. Add PR review time if code review bottlenecks are a concern, and work breakdown (feature vs. bug vs. maintenance split) for stakeholder communication. These five metrics cover delivery speed, quality, and collaboration.
DORA metrics are a specific subset of engineering metrics focused on software delivery performance. Engineering metrics is a broader category that also includes process metrics (PR review time, PR size), team health metrics (work breakdown, sprint velocity), and quality metrics (bug escape rate, test coverage). DORA metrics are a great starting point, but most teams need additional metrics as they mature.
Start by explaining the why: metrics are for improving the system, not evaluating individuals. Make data visible to the whole team, not just managers. Let the team choose which metrics to focus on. And most importantly, act on the data — nothing kills buy-in faster than tracking metrics that never lead to process improvements.
Start with 3-5 metrics tied to specific decisions your team needs to make. Teams that track 15+ metrics typically suffer from dashboard fatigue — the data exists but nobody acts on it. It's better to deeply understand and act on 3 metrics than to passively collect 20.
At minimum, you need a tool that connects to your git repositories and can extract PR data, commit activity, and deployment events. Many teams start with built-in GitHub/GitLab analytics, then move to dedicated tools like Gitmore, LinearB, or Swarmia for automated reporting and deeper analysis. See our git reporting tools comparison for a detailed breakdown.
Gitmore gives engineering teams automated visibility into the metrics that matter most: commit activity, PR throughput, work categorization, and contributor breakdowns. It connects to GitHub, GitLab, and Bitbucket, uses AI to classify every commit and PR into features, bug fixes, refactoring, and maintenance, then delivers structured reports to Slack or email on your schedule.
Instead of building dashboards nobody checks, Gitmore pushes the data to where your team already works. No manual data gathering. No context-switching to a separate analytics tool. Just clear, actionable engineering metrics delivered automatically.
Try Gitmore free — two-minute setup, no credit card, no source code access required.
Explore git reporting for your platform
Automated git reports for your engineering team. Set up in 2 minutes, no credit card required.
Get Started Free