Stack Overflow’s 2025 Developer Survey just dropped, and if you care about developer tools — whether you’re building them, buying them, or using them — the headline number should give you pause: positive sentiment toward AI coding tools has fallen from over 70% in 2023 to just 60% in 2025.
Let me break down the data, because as a data scientist, the numbers tell a more nuanced story than the headline suggests. This isn’t developers rejecting AI — it’s developers getting more realistic about what AI can and can’t do.
The Trust Paradox: Using More, Trusting Less
The adoption numbers are unambiguous: 84% of developers now use or plan to use AI tools in their workflow, up from 76% in 2024. 51% of professional developers use AI tools daily. By any measure, adoption is accelerating.
But trust is moving in the opposite direction:
| Metric | 2024 | 2025 | Change |
|---|---|---|---|
| Positive sentiment | 70%+ | 60% | -10pp |
| Trust AI accuracy | 40% | 29% | -11pp |
| Actively distrust AI | 31% | 46% | +15pp |
| “Highly trust” AI output | ~5% | 3% | -2pp |
That last number is the most telling. Only 3% of developers highly trust what AI tools produce. And among experienced developers (10+ years), the “highly trust” rate drops to 2.6%, while the “highly distrust” rate is 20%.
This isn’t luddism. The people who distrust AI the most are the ones who’ve used it the most and have the deepest understanding of code quality.
The “Almost Right” Problem
The survey identified the #1 developer frustration with AI: 66% cite “solutions that are almost right, but not quite.” Another 45% say debugging AI-generated code is more time-consuming than writing it from scratch would have been.
This matches what I see at Anthropic. The “almost right” problem is arguably harder than “obviously wrong.” When AI generates code that looks correct, passes a cursory review, and maybe even works in simple test cases — but has subtle bugs in edge cases, incorrect assumptions about data types, or security issues that only manifest under load — the time cost of finding and fixing those issues can exceed the time saved by generating the code in the first place.
A Clutch survey of 800 developers found that 59% use AI-generated code they don’t fully understand. That’s not a tooling problem — that’s a professional liability.
The METR Study: Feeling Fast vs Being Fast
This is the data point that should worry the entire industry. A rigorous METR study from July 2025 tracked experienced open-source developers using AI tools on real tasks. The results:
- Developers estimated AI would make them 24% faster
- After the experiment, developers believed they had been 20% faster
- Actual measured performance: 19% slower
The perception gap is massive. Developers feel like AI is helping even when it measurably isn’t. This has significant implications for any organization using “developer satisfaction” or “perceived productivity” as a metric for AI tool ROI.
It also helps explain a counterintuitive finding: senior developers (10+ years) ship 2.5x more AI code than juniors. Not because AI is better for experts — but because experts have the judgment to catch AI mistakes before they ship. AI is a ceiling raiser for people who already know what correct looks like, not a floor raiser for those learning.
What Developers Actually Value
When asked what they value in development tools, developers ranked:
- Reputation for quality — first
- Robust and complete API — second
- AI integration — second to last
Let that sink in. The feature that vendors are pouring billions into is the thing developers care about second to least. Developers want tools that work reliably, not tools that generate code they have to verify.
The Numbers That Aren’t Talked About
- 72% of developers are NOT vibe coding. The narrative that everyone is prompting their way through development doesn’t match the data.
- 52% don’t use AI agents or stick to simpler tools. The agentic future is not here yet for most developers.
- 68% of developers expect employers will mandate AI proficiency. Adoption is increasingly employer-driven, not developer-driven. That’s a significant difference in motivation.
What This Means
I don’t read this data as “AI tools are failing.” I read it as the market maturing. The hype cycle is normalizing into something more sustainable:
- The 60% positive sentiment is probably the real baseline. The 70%+ was novelty-inflated.
- Trust will follow accuracy improvements, not marketing. When tools hallucinate 42% of the time, trust at 29% is rational.
- The best teams are building verification into their workflow, not relying on AI output blindly.
- Developer experience matters more than AI integration. Teams choosing tools based on reliability over AI features will ship more dependable software.
The question for everyone building, buying, or mandating AI tools: are you measuring actual productivity impact, or are you relying on developer perception? Because the METR study suggests those are very different things.
What’s your team’s experience? Has your trust in AI coding tools changed over the past year?