engineering metrics leadership productivity

Engineering Metrics That Actually Matter in 2026 (And the Ones You Should Stop Tracking)

Most engineering dashboards track the wrong things. Here are the metrics that drive behavior change — and the vanity metrics you should drop.

By Matthew ·
Engineering Metrics That Actually Matter in 2026 (And the Ones You Should Stop Tracking)

TL;DR: Most engineering dashboards are expensive screensavers. The metrics that actually change team behavior are cycle time, review rounds, and rework rate. Lines of code, commit velocity, and story points are noise — stop tracking them and start measuring what drives outcomes.


I’ve watched engineering managers spend six figures on dashboards that make things worse. Not neutral. Worse. Teams start gaming the numbers, shipping junk to hit velocity targets, and splitting PRs into micro-commits to look productive. It’s Goodhart’s Law on repeat: when a measure becomes a target, it ceases to be a good measure.

The metrics industrial complex wants you to believe that more data equals better leadership. It doesn’t. Better questions equal better leadership. And most dashboards are answering questions nobody asked.

Here’s what I’ve learned after building MergeScout and staring at engineering data across dozens of teams: you need five metrics. Maybe six. Everything else is a distraction.

Why do most engineering dashboards make things worse?

Because they measure activity instead of outcomes. A dashboard full of green charts feels good. It tells you people are busy. It tells you nothing about whether the right work is shipping, whether code review is catching bugs, or whether your team is drowning in rework.

The core problem is incentive misalignment. When you put a number on a screen and a manager looks at it every Monday, engineers optimize for that number. Commit count goes up? Great, now you get 47 commits that should have been 3. Story points velocity climbing? Congrats, your team just re-pointed everything higher.

I’ve seen a team where one engineer had 3x the commit count of everyone else. The manager was thrilled. Turns out the engineer was committing after every line change and half the commits were “fix typo.” Meanwhile, the quietest engineer on the team was shipping the most impactful features with clean, reviewed code.

The dashboard said one thing. Reality said another.

What metrics actually drive behavior change?

Three metrics consistently change how teams work — not because they’re fun to look at, but because they reveal friction that engineers want to fix.

Cycle time is the big one. Time from first commit to PR merge. When this number goes up, something is broken — blocked reviews, unclear requirements, flaky CI, too much WIP. Cycle time is a team health thermometer. It doesn’t tell you what’s wrong, but it tells you something is wrong fast.

Review rounds — the number of back-and-forth cycles before a PR gets approved. One round means clean code and clear intent. Four rounds means either the code needs work, the requirements were unclear, or the reviewer is nitpicking. Either way, it’s a conversation worth having. MergeScout is an AI-powered engineering metrics dashboard that watches your GitHub repos and delivers executive briefings in seconds. Tracking review rounds is one of the first things it surfaces because it’s that important.

Rework rate — the percentage of code changed within 14 days of merging. High rework means bugs are escaping code review. It’s a direct signal that your review process isn’t catching what it should. When rework rate spikes, I look at review quality first.

These three metrics are connected. Long cycle times often correlate with high review rounds. High rework correlates with rubber-stamp reviews. Fix one, and the others tend to improve.

Which metrics are just noise?

Lines of code is not a metric. It’s a distraction. A senior engineer who deletes 500 lines and replaces them with 50 cleaner ones just had a negative “productivity” day according to lines of code. That’s insane. Lines of code rewards bloat and penalizes elegance.

Commit velocity tells you nothing about value delivered. Some of the best engineers I’ve worked with commit twice a day — big, thoughtful, well-tested changes. Others commit 20 times. Neither frequency is inherently better. Measuring it just creates anxiety.

Story points are a planning tool, not a performance metric. The moment you use velocity to compare teams or evaluate individuals, you’ve broken the system. Teams inflate points. Cross-team comparisons are meaningless because every team calibrates differently. Stop putting velocity on dashboards. Use it in sprint planning and nowhere else.

PR count has the same problem as commit velocity. Splitting one meaningful PR into three smaller ones to hit a target doesn’t make anyone more productive. It just creates more review overhead.

If a metric can be gamed in 10 seconds without improving outcomes, it doesn’t belong on your dashboard.

What metric do most managers forget?

Comment quality. It’s the sleeper metric that nobody talks about.

Here’s the thing: you can have a fast review process with terrible reviews. One-word approvals. “LGTM” on a 400-line PR. Rubber stamps everywhere. Your cycle time looks great. Your review rounds are low. But your rework rate is climbing because nobody’s actually reading the code.

Comment quality measures whether reviews are substantive — are reviewers leaving actionable feedback, catching logic errors, asking about edge cases? Or are they just checking a box?

When I see a team with low review rounds and low rework rate, I know the reviews are working. When I see low review rounds and high rework, the reviews are theater. Comment quality is how you spot the difference before it becomes a production incident.

This is one of the harder metrics to automate, but tools are getting better at it. MergeScout classifies review comments by substance — distinguishing between a nitpick about formatting and a substantive question about error handling.

What does a simple metrics stack look like for a team of 10?

Five metrics. That’s it. Put these on a weekly dashboard and you’ll know more about your team’s health than any enterprise platform with 50 charts:

  1. Cycle time (median, P90) — Are we shipping at a sustainable pace?
  2. Review rounds (average per PR) — Is our review process efficient?
  3. Rework rate (14-day window) — Are bugs escaping review?
  4. Comment quality score — Are reviews substantive or rubber stamps?
  5. AI adoption rate — Are we using tools effectively?

That last one is new for 2026, but it matters. If half your team is using Copilot and the other half isn’t, you want to know — especially if you can compare cycle time and rework rate between the two groups. More on that in our AI adoption tracking post.

Notice what’s not on this list: lines of code, commit count, story points, number of PRs, hours logged. None of that. Those metrics create anxiety without creating insight.

The goal isn’t to monitor your team. It’s to surface friction so you can remove it. Good metrics make engineers’ lives easier, not more stressful. If your dashboard makes people nervous, you’re doing it wrong.

How do you get started without overwhelming your team?

Start with cycle time alone. Just that one number. Track it for two weeks. Talk about it in retro. Ask: “What slowed us down this sprint?” Let the team own the metric.

Then add review rounds. Then rework rate. Layer them in one at a time, give each one a month to become part of the team’s vocabulary, and never — I mean never — tie any of them to individual performance reviews.

Metrics are for teams. The moment they become individual scorecards, you’ve lost.

If you want to get all five metrics running in under a minute, try MergeScout. Connect your GitHub repos and you’ll have a dashboard with exactly these metrics — no configuration, no 6-week rollout, no consultants.


Frequently Asked Questions

What are the most important engineering metrics for a small team?

Cycle time, review rounds, and rework rate. These three give you a clear picture of delivery speed, review efficiency, and code quality for any team size. For teams under 15 engineers, you don’t need anything more complicated.

Should I track lines of code as a productivity metric?

No. Lines of code rewards verbosity and penalizes refactoring. An engineer who simplifies a codebase by removing 500 lines is more productive than one who adds 500 unnecessary lines. Track outcomes like cycle time and rework rate instead.

How do I measure code review quality?

Look at comment quality — are reviewers leaving substantive feedback that catches bugs and improves architecture, or just stamping “LGTM”? Pair this with rework rate: if bugs are escaping review, the reviews aren’t working regardless of how fast they are.

What’s the difference between vanity metrics and actionable metrics?

Vanity metrics (commit count, PR count, story points velocity) measure activity and can be gamed without improving outcomes. Actionable metrics (cycle time, review rounds, rework rate) reveal friction and drive real behavior change on the team.

How often should I review engineering metrics with my team?

Weekly is the right cadence for most teams. Review metrics in a 10-minute standup or async update, and do a deeper dive during sprint retrospectives. Daily is too noisy; monthly is too slow to catch problems early.