Three months ago, our company rolled out GitHub Copilot across all engineering teams. Last week, our VP of Product increased our sprint commitments by 35%. When I pushed back, the response was simple: “You have AI now. The team should be able to handle more.”
Sound familiar?
The Infinite Capacity Myth
I’m seeing this pattern everywhere. Leadership reads the ROI claims, sees “productivity gains,” and assumes teams now have unlimited capacity. Sprint expectations inflate by 30-40%. Roadmap timelines get compressed. Feature requests that would’ve been “Q3 maybe” suddenly become “next sprint definitely.”
But here’s what the data actually shows:
The Productivity Paradox: Research from METR found that developers using AI tools took 19% longer to complete tasks. Yet after the study, those same developers estimated they were 20% faster. We’re not just measuring wrong—we’re feeling wrong about our own productivity.
The Utilization Drop: In our own org, I’ve watched developer utilization of Copilot drop to 22% within 30 days of rollout. The initial excitement fades fast when developers realize the suggestions need as much debugging as writing code from scratch.
The Quality Trade-off: 66% of developers cite inaccurate AI code suggestions as their top challenge. The code looks correct but fails during testing. Time saved in writing gets consumed by checking and editing. Net productivity gain? Minimal at best.
The Burnout Consequence
The real cost isn’t just missed deadlines—it’s people. Since our sprint commitments increased:
- My team leads are working 12-15 hour days trying to meet inflated expectations
- Junior engineers feel like they’re “failing” when they can’t match the supposed AI productivity multiplier
- Our best senior engineer told me she feels like a “janitor cleaning up AI messes” instead of building features
When productivity gains get absorbed by higher demands instead of time savings, burnout follows. And burned-out engineers don’t ship quality software—regardless of what AI tools they have.
The Push-Back Problem
So here’s my question to this community: How do you educate leadership about AI’s actual limitations?
I’ve tried:
- Sharing the research data (eyes glaze over)
- Showing sprint velocity trends (doesn’t match their mental model)
- Explaining that coding is 15% of the job (dismissed as excuses)
What’s worked for you? Do you:
- Track specific metrics that resonate with execs?
- Use particular frameworks for setting realistic expectations?
- Have regular “AI reality check” meetings?
- Frame it differently than “pushing back”?
I’m particularly interested in hearing from other engineering leaders who’ve successfully reset expectations after an AI tool rollout. What data, stories, or frameworks actually got through?
Because right now, the gap between what leadership thinks AI delivers and what teams actually experience is creating unsustainable pressure. And I’m running out of ways to bridge it.
For context: I lead a 40-person engineering org in financial services. We adopted Copilot company-wide in December 2025. Initial excitement was high, but the reality has been much more nuanced than the marketing promised.