25% of AI Investments Deferred to 2027 Amid CFO ROI Demands—While Engineering Still Optimizes for Velocity. Who Defines “Productive” in 2026?
I just came out of a board meeting where our CFO presented data that made me deeply uncomfortable—not because the numbers were wrong, but because they exposed a fundamental misalignment between how engineering measures success and how the business measures value.
The ROI Reckoning Is Here
Forrester estimates 25% of planned AI spend may be deferred into 2027 as enterprises demand to see returns. Meanwhile, 61% of CEOs say they’re under increasing pressure to show returns on AI investments, and only 14% of finance chiefs have seen clear, measurable impact from AI investments.
Our CFO showed me our numbers: We’ve spent $2.3M on AI tooling over the past 18 months—coding assistants, infrastructure automation, ML platforms. Engineering reports we’re shipping 40% faster. Deployment frequency is up 35%. Our DORA metrics look fantastic.
Then she asked: “What revenue did this enable? What costs did it avoid? Which headcount did we defer?”
I didn’t have good answers. We saved time—lots of it—but we didn’t reduce headcount, we didn’t ship features that generated measurable revenue impact, and we didn’t defer hiring because we’re still backfilling positions. The “productivity” vanished into… what exactly? More meetings? Slack threads? Exploration work that didn’t ship?
The Measurement Disconnect
Engineering optimizes for what we can measure: velocity, cycle time, deployment frequency, change failure rate. These are the DX Core 4 metrics we’ve all adopted. But our CFO pointed out something uncomfortable: these metrics measure activity, not outcomes.
She showed me her framework for evaluating AI ROI:
- Time saved that converts to headcount avoided (we hired the same number of people)
- Faster time-to-market that captured revenue opportunity (our product velocity didn’t change)
- Cost reduction through automation (our AWS bill went up, not down)
- Quality improvement that reduced support costs (our incident rate is flat)
By her math, we spent $2.3M and generated… maybe $200K in measurable value? An 8.7% ROI. She’s deferring our next round of AI tool investments until we can articulate clearer value capture.
Who Defines “Productive”?
Here’s what keeps me up: Are we measuring the wrong things, or are we actually not being productive in the way that matters to the business?
CFOs need to move away from traditional financial metrics for AI because different AI investments generate value differently. But the inverse is also true: engineering needs to move away from purely velocity-based metrics when evaluating productivity gains.
The uncomfortable question: If we’re “40% more productive” but the business can’t point to $920K in captured value (40% of our $2.3M spend), are we actually productive or just… busy?
The Translation Problem
I think the core issue is a translation failure. Engineering speaks in time saved and features shipped. Finance speaks in revenue enabled and costs avoided. The job is to translate engineering metrics into outcomes a CFO can defend.
Some teams are doing this well. One framework maps change lead time to revenue velocity, change failure rate to incident cost, and deployment recovery time to downtime cost. But this requires instrumenting the entire delivery pipeline to connect technical activity to business impact—work most engineering teams haven’t done.
What I’m Changing
Starting next quarter, I’m requiring every AI tool investment to articulate its value hypothesis:
- For productivity tools: What work will this eliminate, and how will we redeploy that capacity to measurable outcomes?
- For quality tools: What incidents will this prevent, and what’s the cost of those incidents?
- For velocity tools: What time-sensitive market opportunities will faster shipping enable us to capture?
We’re also changing how we report engineering productivity. Instead of “deployment frequency up 35%”, we’re tracking “features shipped that drove measurable engagement lift” and “technical debt resolved that reduced on-call burden.”
The Question
How do you bridge the CFO-CTO measurement gap?
Are there engineering teams successfully translating technical productivity into financial outcomes? Or is this fundamentally hard because much of engineering value is enabling and indirect—harder to measure but still real?
I’m particularly curious: If 25% of AI investments are being deferred, does that mean companies are cutting genuinely productive tooling, or are they finally holding engineering accountable for value capture? Or both?
Because right now, my CFO is questioning whether “shipping 40% faster” matters if the business outcomes stay flat—and I’m starting to think she’s right.