Two-Thirds of Developers Don't Trust Your Productivity Metrics. DORA vs SPACE vs DX Core 4—Which Framework When Your Exec Wants "One Number"?

Last Tuesday, I sat in a board meeting where our CFO looked me straight in the eye and asked: “Michelle, can you just give me ONE number that tells me if engineering is productive?”

I wanted to laugh. Then cry. Then pull up the 47 tabs I have open comparing DORA vs SPACE vs DX Core 4.

Here’s what I said instead: “If I gave you one number, it would be wrong. And 66% of my team wouldn’t trust it anyway.”

The Trust Crisis Nobody’s Talking About

According to JetBrains’ 2025 State of Developer Ecosystem survey, 66% of developers don’t believe current metrics reflect their true contributions. That’s not a measurement problem—that’s a legitimacy crisis.

And it gets worse. 55% of developers reported that metrics were used punitively—to justify layoffs, enforce micromanagement, or penalize teams during performance reviews. When your measurement system becomes a weapon, developers learn to game it or ignore it entirely.

Meanwhile, 29.6% of platform teams measure nothing at all, making ROI conversations with executives impossible. We’re stuck between metric distrust and metric vacuum.

Framework Fatigue: DORA, SPACE, DX Core 4… What’s the Difference?

If you’re like me, you’ve spent the past 6 months drowning in framework comparisons:

DORA (DevOps Research and Assessment): The incumbent. Four metrics—Deployment Frequency, Lead Time for Changes, Change Failure Rate, and Mean Time to Recovery. DORA is focused on CI/CD pipeline health, which matters, but it’s not the full picture. You can have stellar DORA metrics while shipping features nobody uses.

SPACE Framework: The academic darling. Five dimensions—Satisfaction, Performance, Activity, Communication & collaboration, Efficiency & flow. SPACE explicitly acknowledges “there is no one measure of productivity” and forces you to think multidimensionally. But try presenting that to a board that wants simplicity.

DX Core 4: The new unifier. Four dimensions—Speed, Effectiveness, Quality, Business Impact. Developed in collaboration with the authors of DORA, SPACE, and DevEx, DX Core 4 aims to be practical: can be deployed in weeks (not months), uses readily-available system metrics, and minimizes survey fatigue.

The Executive Dilemma: They Want Simplicity, Research Says “No Single Metric”

Here’s the tension: Every framework’s documentation explicitly warns against single-number productivity measurement. The SPACE framework’s core insight is that multidimensional measurement is non-negotiable.

But executives operate in a world of comparable simplicity:

  • Sales has ARR and win rate
  • Marketing has CAC and LTV
  • Finance has margins and burn rate
  • Engineering has… “it’s complicated”?

When your board member asks for “one number,” they’re not being stupid. They’re pattern-matching from every other business function they oversee. Engineering feels like a black box.

What We Actually Use (And Why It’s Still Imperfect)

At my SaaS company, we’ve landed on a hybrid approach:

  1. DORA metrics for pipeline health: Deployment frequency and lead time give us operational baseline
  2. SPACE satisfaction surveys (quarterly): Developer sentiment is our early warning system for attrition and burnout
  3. Custom “business impact” metric: Features shipped weighted by actual revenue impact or user adoption (tracked in Amplitude)

We present this as an “Engineering Health Dashboard” to leadership—three sections, not one number. It took 6 months of education, but now our CFO understands that asking for “one number” is like asking “what’s one number that captures your company’s financial health?” (Hint: it’s never just revenue or just profit.)

But I’ll be honest: It’s still imperfect. Platform team work doesn’t map cleanly to revenue. Incident prevention efforts show up as “no deploys this sprint” in DORA. Developer happiness surveys have lag time—by the time satisfaction drops, your best engineers are already interviewing elsewhere.

The Hard Truth: You Can’t Collapse Complexity, But You CAN Tell a Story

What I’ve learned: Modern best practice is combining DORA with SPACE, DX Core 4, and flow metrics to capture the full picture. But executives don’t read 30-page reports.

The breakthrough wasn’t simplifying the data. It was improving the narrative.

Now our monthly engineering review is 4 slides:

  • Slide 1: Speed (DORA deployment frequency + lead time)
  • Slide 2: Quality (DORA change failure rate + P0 incidents)
  • Slide 3: Team health (SPACE satisfaction + retention rate)
  • Slide 4: Impact (Revenue-enabled features + platform adoption)

Each slide includes 1-2 anecdotes. Numbers tell you what’s happening. Stories tell you why it matters.

My Question for This Community

How do you balance executive simplicity needs with measurement complexity?

  • What framework(s) have you adopted and why?
  • How do you handle the “just give me one number” request?
  • For those with platform teams—how do you measure value when there’s no direct revenue attribution?
  • Have you found ways to rebuild developer trust in metrics after years of punitive use?

I’m especially curious to hear from product and business leaders: What would help you understand engineering productivity without forcing oversimplification?

Because right now, we’re caught between executive pressure for simplicity and developer distrust of any measurement at all. And with AI coding tools reshaping how we work, 2026 is forcing a reckoning on what “productivity” even means.

What’s working for you?

This resonates hard, Michelle. Our CFO asked me the exact same question last quarter: “Luis, what’s the ONE metric I should watch?” I think every engineering leader has lived this moment.

The Regulated Environment Adds Another Layer

In financial services, we have an additional challenge: our metrics also need to satisfy audit and compliance requirements. Regulators want to see evidence of “appropriate controls” over our systems. Try explaining SPACE framework dimensions to a bank examiner who wants proof your deployment process is “adequately controlled.” :sweat_smile:

We ended up needing metrics that serve three masters:

  1. Engineering teams (to improve our processes)
  2. Business executives (to demonstrate value and ROI)
  3. Compliance and audit (to prove control and risk management)

Our Journey: DORA → Developer Revolt → Hybrid

We started with pure DORA metrics about 18 months ago. Seemed sensible—four clear numbers, well-researched, widely adopted.

Three months in, one of my senior engineers pulled me aside after a 1:1: “Luis, I just spent 2 weeks on incident prevention work—refactoring our alerting system, improving error handling, documenting failure modes. Zero deploys. Zero features shipped. Am I unproductive according to your dashboard?”

That hit me like a truck. We were measuring activity that’s visible in the pipeline but ignoring crucial work that prevents fires before they start.

What We Use Now

We’ve evolved to a hybrid approach that my CFO grudgingly accepts and my team actually trusts:

1. DORA metrics (for pipeline health)

  • Deployment frequency, lead time, MTTR
  • But we contextualize them: “Platform team deployed 3 times this sprint—here’s why that was the right cadence for their work”

2. Developer sentiment surveys (quarterly)

  • Borrowed from SPACE framework
  • 10-question pulse: satisfaction, tools quality, collaboration effectiveness
  • Anonymous, but we slice results by team to identify problems
  • This became our early warning system for burnout and attrition

3. “Business outcomes” tracking

  • Features shipped weighted by actual adoption (DAU/MAU)
  • Critical bugs fixed (P0/P1)
  • Technical debt reduction (measured as story points, I know, not perfect)
  • Compliance and security work (tracked separately because it’s non-negotiable)

The Dashboard as Narrative Device

I stole your approach, Michelle—instead of trying to compress this into “one number,” we present an “Engineering Health Dashboard” every month.

Four sections:

  • Delivery health (DORA metrics + context)
  • Team health (sentiment surveys + retention rate)
  • Business impact (features adopted + critical bugs fixed)
  • Technical foundation (platform work + security/compliance)

Each section includes 2-3 sentences explaining what changed and why it matters. The CFO reads the narrative, glances at the numbers, and asks 1-2 questions. Total time: 15 minutes in our monthly business review.

It took about 4 months of education and a few tough conversations (“This isn’t Netflix’s blog post summary—every company is different”), but now leadership gets it.

My Question Back to the Community

How do you handle metrics for platform teams vs product teams?

Platform teams have fundamentally different value delivery models:

  • Product teams → features → users → revenue (relatively direct line)
  • Platform teams → capabilities → developer productivity → features → users → revenue (very indirect)

We’re still struggling with this. Right now we track “adoption rate” (what % of product teams use the new CI/CD pipeline, for example), but it feels incomplete. Platform work is about enablement, reliability, and reducing cognitive load—hard to quantify.

Anyone cracked this nut? What works for measuring platform/infrastructure team impact in a way that both developers and executives trust?

Michelle, this is SO on point. And Luis, your senior engineer’s feedback about incident prevention work being “invisible” in DORA metrics—I’ve heard that exact complaint multiple times.

Reframing the “One Number” Request

Here’s what I’ve learned after being asked this question by three different CEOs across Google, Slack, and my current EdTech startup:

When an exec asks for “one number,” they’re not actually asking for one number.

They’re asking: “Help me understand if we’re winning.”

They want confidence that engineering is delivering value, that we’re not wasting money, that we’re competitive with other companies in our space. “One number” is executive shorthand for “I’m overwhelmed by 47 tabs of metrics research and need you to synthesize this for me.”

Our job isn’t to give them one number. It’s to translate engineering health into terms that make sense in the boardroom.

My Experience: From JIRA Velocity to SPACE Framework

At Slack, we went through this exact journey around 2022-2023.

Phase 1: JIRA velocity hell

  • Engineering managers tracking story points completed per sprint
  • Quarterly reviews where leaders compared teams: “Why is Team A at 45 velocity and Team B at 28?”
  • Result: Teams started inflating story points to look more productive
  • Developer trust in metrics: basically zero

Phase 2: DORA metrics

  • Big improvement! Focused on outcomes (deployed code) not outputs (estimated effort)
  • But… similar problem to Luis’s experience. Infrastructure work, security work, research spikes—all invisible or penalized

Phase 3: SPACE framework

  • Introduced in partnership with Dr. Nicole Forsgren (one of DORA’s creators!)
  • Five dimensions: Satisfaction, Performance, Activity, Communication, Efficiency
  • Explicitly multidimensional—acknowledged “there is no one measure of productivity”

Leadership reaction: “This is too complicated. Can you simplify it?”

Sound familiar? :sweat_smile:

The Breakthrough: Executive Summary + Deep Dive Structure

Here’s what finally worked:

Monthly “Engineering Health Review” with execs—two formats:

1. Executive Summary (4 slides, 10 minutes)

  • Slide 1: Speed (how fast are we shipping?)
  • Slide 2: Effectiveness (are we shipping the right things?)
  • Slide 3: Quality (is what we shipped working?)
  • Slide 4: Impact (did it matter to users/business?)

Each slide has:

  • 1-2 key metrics (DORA + SPACE + custom)
  • Trend arrow (↑↓↔) compared to last quarter
  • 2-3 sentences of narrative context
  • One specific anecdote/example

2. Deep Dive (available on request, 25 slides)

  • Full SPACE framework breakdown
  • Per-team metrics
  • Survey results with demographic cuts
  • Technical debt tracking
  • Incident analysis

Leadership reads the 4-slide version. The 25-slide version exists to answer “show me the details” questions and to give the team confidence we’re measuring the right things comprehensively.

The Trust Turnaround

Here’s the stat that matters: We ran internal surveys about metric trust.

Before SPACE framework: 35% of engineers trusted that metrics reflected their work accurately
After 6 months with SPACE: 78% trust

What changed?

  1. Involved developers in metric selection - We ran workshops where teams voted on which SPACE dimensions mattered most for their work
  2. Made it clear metrics were for IMPROVEMENT not PUNISHMENT - Explicitly told managers: “These are for identifying blockers, not performance reviews”
  3. Showed we listened to feedback - When engineers said “Activity metrics are bullshit,” we deprioritized LOC and PR counts, emphasized satisfaction and flow
  4. Closed the loop - When metrics showed problems (satisfaction dropping, communication breaking down), we took visible action to fix it

Trust isn’t about choosing the “right” framework. It’s about demonstrating that measurement leads to improvement, not judgment.

My Advice: Don’t Fight the “One Number” Request

Treat it as an opportunity to educate.

When a CEO or CFO says “just give me one number,” I now say:

“I hear you wanting simplicity. Here’s what I can give you: a quarterly Engineering Health Score that combines four dimensions into a 0-100 rating with color coding (red/yellow/green). But if that number drops, you’ll need to look at the detail to understand WHY. Is it a speed problem? Quality problem? Team morale problem? Each has a different solution. Can I show you how this works?”

9 times out of 10, they say yes to the deeper explanation once you acknowledge their need for executive-level simplicity.

Question for Product Leaders Like @product_david

What would help YOU understand engineering productivity better without forcing oversimplification?

I’m genuinely curious: When you’re in sprint planning or roadmap discussions, what signals from engineering make you confident vs worried about execution capacity? What would you want in an ideal “Engineering Health” dashboard that would help you make better product decisions?

Because ultimately, metrics should enable better cross-functional collaboration—not just satisfy executive reporting requirements.