I just got off a board call where our CFO announced we’re cutting 25% of our AI tool budget. The reason? “We can’t measure the ROI.”
Here’s what’s keeping me up at night: We’re measuring the wrong things.
The Numbers Don’t Lie (But They Don’t Tell the Whole Story Either)
A recent CFO survey found that only 14% of finance leaders see clear, measurable impact from AI investments. Meanwhile, 61% of CEOs are under pressure to show returns. The disconnect is real, and it’s costing us.
But here’s the paradox that should concern every engineering leader: Developers say they’re 20-40% more productive with AI coding assistants. Yet companies with high AI adoption aren’t shipping faster or more reliably.
In fact, one study I reviewed last week found that developers using AI tools actually took 19% longer to complete tasks than without—even though they estimated they were 20% faster. We’re not just bad at measuring AI impact. We’re systematically wrong about what’s happening.
We’re Optimizing for the Wrong Metrics
The problem isn’t AI. It’s that we’re measuring individual output instead of organizational outcomes.
CFOs want to see:
- Cost per feature
- Lines of code per engineer
- Time to close tickets
But these metrics miss what actually matters:
- Time to onboard new engineers (we’ve cut this nearly in half with AI pair programming)
- Quality of architectural decisions (engineers explore more options, make better tradeoffs)
- System reliability improvements (AI helps catch edge cases we used to miss)
- Competitive positioning (we can tackle problems that would’ve required 2x the headcount)
The hardest conversation I’ve had recently was with our CFO about why our “productivity” metrics look flat despite every engineer loving our AI tools. The answer: The bottleneck moved. We’re not blocked on writing code anymore. We’re blocked on code review, integration testing, and cross-functional alignment.
AI didn’t make those problems worse—it just revealed them as the actual constraints.
What’s Actually Working
Organizations seeing real ROI from AI share three things:
- Baseline measurements before rollout - You can’t measure improvement if you don’t know where you started
- Workflow redesign before tool adoption - MIT, McKinsey, and Wharton research all say the same thing: transformation fails when treated as a technology rollout
- Alignment on what success means - 65% of orgs lack agreement between CFO, CTO, and business leaders on how to measure AI success
At my previous company, we treated AI tool adoption like an ERP decision: clear business case, defined success metrics, post-implementation review. It worked. Here, we let teams pick their own tools and hoped for bottom-up ROI. It didn’t.
The Question I’m Wrestling With
How do we measure things that matter to both engineering excellence and financial accountability?
I’m starting to think the answer is: Stop measuring productivity. Start measuring capability.
Can we solve problems we couldn’t solve before? Can smaller teams tackle bigger challenges? Are we making better decisions faster? Are we building more defensible competitive advantages?
These are harder to quantify. But they’re also what boards actually care about when they ask “What are we getting for our AI investment?”
What metrics are you using to measure AI impact? What’s working? What conversations are you having with your CFO?
I’d love to hear how other technical leaders are navigating this. Especially if you’ve found frameworks that resonate with finance teams.