I’ve been VP of Finance at our Series B fintech startup for three years. Yesterday’s Q1 budget review was a wake-up call.
Our CFO pulled up our AI spending: GitHub Copilot for 50 engineers, a customer service AI platform, and an ML recommendation engine. Total annual cost: $430K. His question stopped the room: “Show me the unit economics.”
We couldn’t. Not convincingly.
The Industry-Wide Reality Check
We’re not alone in this struggle. Forrester’s 2026 research reveals that only 14% of CFOs report seeing clear, measurable impact from their AI investments. That means 86% of companies are spending significant money without being able to prove the return.
The market is course-correcting. Forrester predicts enterprises will defer 25% of planned AI spend to 2027. This isn’t AI failing—it’s accountability arriving.
The Fundamental Shift Happening Right Now
In 2024-2025, AI spending came from “innovation budgets” with loose ROI requirements. Experimentation was the priority. “Move fast and learn” was the mantra.
In 2026? AI spending is moving into operational technology budgets. Our CFO literally said: “We evaluate AI investments with the same rigor as ERP implementations or headcount decisions.”
This is the accountability era for AI. And frankly, it’s necessary and overdue.
What Actually Passed CFO Scrutiny
After that budget meeting, I spent two weeks building a proper ROI analysis for our GitHub Copilot deployment. Here’s what convinced our CFO to renew:
Cost Analysis:
- $19/month/developer × 50 engineers × 12 months = $11,400 annual
- Integration, training, and support time: ~$8,000 one-time
- Total year 1 investment: $19,400
Measured Benefits:
- 18% reduction in time-to-PR completion (measured over 3-month instrumented pilot)
- Translates to ~0.9 hours saved per developer per week
- 50 developers × 0.9 hrs × 48 weeks × $75/hour blended rate = $162,000 annual productivity value
- Avoided 1 contractor hire ($120K annually) due to increased team throughput
Bottom line ROI: 8.3x in year 1, even using conservative assumptions.
The critical difference was rigorous measurement. We instrumented the pilot environment, tracked time-to-PR before and after deployment, surveyed developers weekly on perceived productivity, and analyzed PR complexity and volume changes.
The Tension I’m Wrestling With
Here’s what keeps me up at night: Recent research shows 61% of business leaders feel more pressure to prove AI ROI compared to a year ago. That pressure drives necessary financial discipline.
But are we inadvertently killing transformative innovation? The most valuable AI applications might require 12-18 months to demonstrate full ROI. If finance teams demand quarterly proof points, do we lose the opportunities that could fundamentally change our businesses?
I don’t have a clean answer. I’m trying to balance rigorous accountability with the breathing room our engineering and product teams need to discover the next generation of AI value.
For others navigating similar conversations with finance stakeholders: What frameworks are you using to justify AI investments? Which metrics actually move the needle? How do you balance short-term accountability with longer-term innovation horizons?
I’d genuinely love to learn how other organizations are managing this shift from experimentation to accountability.