I need to talk about something that the productivity-focused AI coding discussions consistently underweight: the security implications of review fatigue on AI-generated code. Because from a security perspective, we are building a catastrophe.
Here is the threat model in plain terms: AI coding tools dramatically increase code volume. Review capacity stays flat. Reviewers get fatigued and start rubber-stamping. Attackers know this, and they will exploit it.
This is not theoretical. It is already happening in open source, and it is a matter of time before it becomes a widespread problem in internal codebases too.
The Attack Surface Nobody Is Modeling
Let me describe the attack surface that AI-accelerated review fatigue creates:
1. Volume Camouflage
When a team produces 15 PRs per week, a malicious PR stands out. When a team produces 50 PRs per week, the signal-to-noise ratio drops dramatically. A reviewer who is fatigued from reviewing 10 PRs that morning is more likely to miss a subtle vulnerability in PR number 11. Attackers have always relied on overwhelming defenders – AI code volume does this automatically.
2. AI Pattern Exploitation
AI-generated code creates recognizable patterns. Reviewers learn to expect these patterns and develop “AI code blindness” – they start skimming AI code because it “usually looks right.” This is exactly the opening for an attacker (or a compromised AI model) to introduce malicious code that hides behind the expected AI style.
A recent study found that 45% of AI-generated code contains security vulnerabilities. Not all of these are exploitable, but the volume means even a small percentage of exploitable vulns translates to a significant attack surface.
3. Supply Chain Contamination
This connects to the slopsquatting problem. AI tools sometimes hallucinate package names – 20% of AI-generated code references non-existent packages, and 43% of hallucinated names recur consistently. Attackers register these hallucinated package names and upload malicious code. A developer using AI to generate code pulls in a dependency that the AI invented, and the attacker was waiting for it.
In a world where reviewers are overwhelmed, the chance of someone catching “wait, this package did not exist last week” drops significantly.
4. The Insider Threat Multiplier
Consider a scenario where a malicious insider uses AI to generate a large volume of legitimate-looking PRs to build trust and create review fatigue among their colleagues. After weeks of AI-generated code that passes review easily, they slip in a PR with a subtle backdoor. The reviewers, conditioned to approve their AI-generated output quickly, miss it.
This is not far-fetched. It is a straightforward adaptation of existing insider threat tactics to the new AI-code-generation reality.
What the Data Says About Security Review Quality
GitHub is literally building a “kill switch” for pull requests because AI-generated contributions are overwhelming open-source maintainers. In the enterprise space, SonarSource’s 2026 developer survey found that 95% of developers spend effort reviewing and correcting AI output, but 59% rate that effort as only “moderate” – which in security terms means they are probably missing things.
The Cortex 2026 benchmark shows incidents per PR up 23.5% alongside higher AI adoption. Some portion of those incidents are almost certainly security-related. And change failure rates are up 30%, which means more emergency patches, more rushed fixes, more opportunities for the fix-for-the-fix to introduce new vulnerabilities.
My Actual Recommendations
I spend a lot of time doing threat modeling for organizations, and here is what I am recommending to every client right now:
1. Separate security review from code review.
Do not expect the same developer who reviews business logic to also catch security issues in AI-generated code. Have dedicated security review for any PR that touches authentication, authorization, data handling, or external interfaces – regardless of whether it is AI-generated or not.
2. Mandatory SAST/DAST before human review.
No AI-generated PR should reach a human reviewer without first passing a comprehensive static application security testing (SAST) scan. This is table stakes. If your pipeline does not do this, fix it today.
3. Dependency lockfiles and package verification.
Every new dependency introduced by AI should be verified against known registries. Automated checks should flag any dependency that did not exist 30 days ago or has fewer than a threshold number of downloads.
4. AI code provenance tracking.
You need to know which code in your codebase was AI-generated. Not for blame, but for risk assessment. AI-generated code has a different risk profile than human code, and your security posture should reflect that.
5. Slow down on sensitive paths.
For code that handles money, personal data, authentication, or external API interactions: reject the pressure to review faster. These paths deserve the same review depth they always did, regardless of how the code was generated. If that means those PRs take longer, that is the correct trade-off.
The Uncomfortable Parallel
I keep thinking about the early days of DevOps, when the mantra was “ship faster, break things, fix forward.” Security teams warned that faster deployment without proportional security investment would create problems. Those warnings were largely ignored until we got SolarWinds, Log4Shell, and the continuous parade of supply chain attacks that followed.
We are at the same inflection point with AI-generated code. The industry is focused on productivity acceleration and treating review capacity as an inconvenience to be optimized away. Security teams are warning that degraded review quality creates exploitable gaps. The question is whether we listen this time, or wait for the inevitable major incident.
What are you seeing in your organizations? Are security reviews getting the same attention they did before AI coding tools? Or is the pressure to ship faster winning?