The CFO AI skepticism conversation has me thinking about a fundamental problem: we’re terrible at measuring AI value in ways that finance actually cares about.
When I talk to our CFO about AI investments, the conversation usually goes like this:
Me: “This AI initiative will improve developer productivity.”
CFO: “By how much? And how does that translate to revenue or cost savings?”
Me: “Well, it’s hard to measure precisely, but engineers report feeling more productive…”
CFO: skeptical silence
We’re losing that conversation because our metrics are weak. “Developer satisfaction improved” or “code quality scores went up” don’t translate to business value in finance’s language. And increasingly, I think that’s on us as technology leaders, not on finance for being unreasonable.
The Measurement Gap:
Here are the metrics we typically use to justify AI investments:
- Lines of code written (meaningless, often inverse of value)
- Developer productivity surveys (subjective, hard to tie to outcomes)
- Technical metrics like model accuracy (doesn’t translate to business impact)
- “Estimated time saved” (usually inflated, rarely validated)
- Anecdotal success stories (not scalable or systematic)
Here are the metrics finance cares about:
- Revenue impact (new revenue generated or revenue protected)
- Cost reduction (actual headcount avoided or expenses eliminated)
- Time to market (shipping features faster that drive business results)
- Risk mitigation (security, compliance, operational risks reduced)
- Customer impact (retention, satisfaction, expansion that ties to revenue)
There’s a translation layer we’re missing. And in the current CFO skepticism environment, that missing translation is costing us credibility and budget.
The Attribution Problem:
Even when we try to measure business impact, AI investments have an attribution problem:
- If we ship a feature faster using AI coding assistants, how much of that speed was AI vs just having experienced engineers?
- If revenue goes up after launching an AI-powered feature, how much was the AI vs other factors like marketing, pricing, market conditions?
- If costs go down after implementing AI automation, how much would have gone down anyway from other efficiency efforts?
We need frameworks for rigorous attribution, not just correlation. But I haven’t seen many good examples of companies doing this well.
What I’m Trying:
I’ve started requiring every significant AI investment to have a “business value hypothesis” upfront:
- Revenue hypothesis: “This AI feature will increase conversion by X% based on A/B tests, generating $Y in new ARR”
- Cost hypothesis: “This AI automation will reduce support tickets by X%, avoiding $Y in support headcount”
- Time hypothesis: “This AI tool will reduce feature development time by X%, allowing us to ship Y more revenue-generating features per quarter”
Then we actually measure against those hypotheses post-launch. Not perfect, but it’s forcing more discipline.
The Compound Value Problem:
Here’s where it gets tricky: some AI investments have value that compounds over time in ways that are hard to measure upfront.
Example: We built an ML platform that makes it easier for product teams to ship AI features. The first feature took 6 months and cost $500K. The second feature took 3 months and cost $200K. The third feature took 1 month and cost $50K.
How do you measure the ROI of that platform investment? Traditional ROI calculations might say “it took 18 months to break even.” But the compounding value means every subsequent AI feature is dramatically cheaper and faster.
Finance doesn’t have great frameworks for valuing that kind of platform investment, and we haven’t given them better frameworks.
The Risk/Opportunity Cost Question:
Another measurement challenge: how do we value defensive AI investments—spending that doesn’t create new value but protects existing value?
- Investing in AI security to prevent future breaches
- Building AI monitoring to catch production issues faster
- Implementing AI quality checks to reduce customer churn from bugs
These are “insurance” investments where the ROI is “bad things that didn’t happen.” Finance struggles to value these, and honestly, so do we.
Similarly, what’s the cost of NOT investing in AI? If competitors ship AI features and we don’t, we might lose market share—but that’s a counterfactual that’s impossible to measure precisely.
What I’m Looking For:
I need frameworks and metrics that:
- Translate technical improvements to business outcomes in ways finance can model and believe
- Handle attribution rigorously so we’re not claiming credit for results we didn’t drive
- Value long-term/compound benefits not just immediate ROI
- Account for risk mitigation and opportunity cost not just direct value creation
- Work at different scales from small experiments to large platform investments
Questions for the community:
- What metrics are you using to measure AI investment value that actually resonate with your CFO?
- How do you handle the attribution problem when multiple factors contribute to results?
- What frameworks exist for valuing platform/infrastructure AI investments with compound benefits?
- How do you measure defensive AI investments where the value is “bad outcomes prevented”?
- Are there examples of companies doing AI value measurement really well that we can learn from?
The CFO skepticism is a symptom. The underlying disease is our inability to measure and communicate AI value in business terms. We need to get better at this, fast.