I’m preparing our Series C fundraising materials, and investors keep asking the same question: “What’s your AI infrastructure cost as a percentage of revenue?” I genuinely don’t have a good answer, and it’s making me question whether we understand our own unit economics.
AI Cost Management is Now Priority #1
According to the latest State of FinOps 2026 report, AI cost management is the #1 desired skillset across organizations of all sizes - 98% adoption, up from just 63% last year. That’s a massive shift.
And I understand why. AI costs are eating our cloud budget.
Our cloud spend breakdown (2024 vs 2026):
2024:
- Compute: K/month (65%)
- Storage: K/month (17%)
- Data transfer: K/month (12%)
- Other: K/month (6%)
- Total: K/month
2026:
- Traditional compute: K/month (29%)
- AI/ML compute (GPUs): K/month (52%)
- Storage: K/month (11%)
- Data transfer: K/month (7%)
- Other: K/month (1%)
- Total: K/month
AI infrastructure went from 0% to 52% of our cloud spend in 18 months. And it’s wildly unpredictable.
The Forecasting Nightmare
Traditional infrastructure is predictable:
- “We need 10 servers to handle 100K users”
- “Each server costs /month”
- “If we grow to 150K users, we need 15 servers”
- Math works out, forecasting is straightforward
AI infrastructure is chaos:
- GPU spot instance prices swing 10x based on availability
- Model inference costs depend on request complexity (token count, image size)
- Training runs have unpredictable duration (convergence isn’t linear)
- New models launch monthly with different cost/performance tradeoffs
Example from last month:
We budgeted K for GPU compute. Actual spend: K.
Why?
- Spot instance availability dropped mid-month (AWS announced new AI services)
- Fell back to on-demand instances at 3x cost
- Product launched a new feature with longer AI prompts (more tokens = higher cost)
- Engineering experimented with larger model (needed more GPU memory)
How do I explain this variance to our CFO? To potential Series C investors?
FinOps Is Now an Engineering Problem
The State of FinOps report shows 78% of FinOps teams now report to CTO/CIO, up from 60% in 2023. This isn’t a finance function anymore - it’s a technical architecture decision.
And that makes sense. The biggest cost optimization opportunities aren’t in vendor negotiation - they’re in:
- Which AI model to use (GPT-4 vs Claude vs open-source)
- Model quantization and optimization techniques
- Batch processing vs real-time inference
- GPU instance type selection and scheduling
- Prompt engineering to reduce token usage
These are engineering decisions with massive financial implications. But my engineering team doesn’t think about costs, and my finance team doesn’t understand AI architecture.
FinOps Expanding Beyond Cloud
Here’s what’s making my job even harder: FinOps is no longer just cloud costs.
According to the 2026 report:
- 90% managing SaaS costs (up from 65%)
- 64% managing software licensing (up 15%)
- 57% managing private cloud (up 18%)
- 48% managing data centers (up 12%)
We’re supposed to be “cloud finance” experts, but now I’m tracking:
- AWS/GCP cloud spend
- OpenAI API costs
- Anthropic API costs
- Hugging Face inference endpoints
- GitHub Copilot seats
- Datadog subscription
- SaaS tools across the company
- Software licenses for development tools
I need a unified view of all technology spending, but every vendor has different billing models, different APIs, different reporting.
The Questions I’m Struggling With
1. How do I budget AI costs when they’re fundamentally unpredictable?
Should I:
- Use historical average + 50% buffer? (Feels arbitrary)
- Model per-request costs and multiply by traffic forecast? (Traffic forecasts are also wrong)
- Just tell the board “AI costs will fluctuate 30-40%”? (They won’t love this)
2. How do I align engineering incentives with cost optimization?
Engineers want the best model for the best user experience. Finance wants predictable, optimized costs. These often conflict.
Do I:
- Make cost metrics part of engineering performance reviews? (Feels wrong)
- Create cost budgets per team? (Creates friction and gaming)
- Just accept that this is a business cost of AI-driven products? (CFO says no)
3. What FinOps tools actually work for AI costs?
We’ve tried:
- AWS Cost Explorer (useless for AI-specific analysis)
- Custom dashboards (maintenance burden)
- Third-party FinOps tools (expensive, don’t handle AI well)
Are there AI-specific cost management tools that actually deliver value? Or is this still too nascent?
4. How do other companies model AI cost in unit economics?
For our Series C pitch, investors want to see:
- Cost per user
- Gross margin
- Path to profitability
But AI costs break traditional unit economics:
- One power user might cost 100x more than average user (based on AI usage)
- We can’t easily attribute AI costs to specific customers
- Costs depend on model choice, which changes quarterly
How are other AI-driven companies modeling this for investors?
Looking for Perspectives
I’d love to hear from:
-
Engineering leaders: How do you think about AI costs in architecture decisions? Do you have cost budgets? How do you balance performance vs cost?
-
Other finance folks: How are you forecasting AI costs? What models or frameworks work? How do you present this to boards and investors?
-
FinOps practitioners: What tools and practices actually help with AI cost management?
This feels like a 2026-specific problem where best practices are still being written. Help me learn from your experiences.