AI Has This “Overwhelming Tendency” to Ignore Our Repo Conventions—How Are You Enforcing Consistency at Scale?
Three weeks ago, I reviewed a PR where our AI coding assistant generated a beautiful, working feature… in completely the wrong architectural style. The code worked. Tests passed. But it ignored every convention we’d built over two years of microservices evolution.
The problem is getting real. 93% of developers are using AI coding tools, and 41% of all code in 2026 is AI-generated. That’s not a pilot anymore—that’s production at scale. But here’s what nobody’s talking about: AI has this overwhelming tendency to not understand what the existing conventions are within a repository.
What We’re Seeing in Our Financial Services Codebase
I lead a team of 40+ engineers maintaining critical banking infrastructure. Here’s what architectural drift looks like in practice:
Naming inconsistency: AI toggles between camelCase and snake_case based on… vibes? One service uses getUserBalance, the next uses get_account_balance. Grep becomes useless. Refactoring becomes dangerous.
Architecture bypass: We have battle-tested middleware for transaction logging, rate limiting, and compliance auditing. AI-generated services wire around it because the agent doesn’t understand why those layers exist. Now we have compliance gaps that only surface during audits.
Documentation drift: Our team wiki says “all REST endpoints must include OpenAPI specs.” AI generates perfect implementations with zero documentation. Six months later, nobody remembers what those endpoints do.
The statistics validate what we’re experiencing: 96% of developers don’t fully trust the functional accuracy of AI-generated code. We’re all working with tools we fundamentally don’t trust to follow our rules.
Why Documentation Alone Doesn’t Work
We tried the standard approach: CODING_STANDARDS.md, ARCHITECTURE.md, AGENTS.md explaining our patterns with examples. The AI reads them. Sometimes follows them. Often… invents its own interpretation.
Here’s the core issue: Documentation files are fundamentally suggestions, not guarantees. AI can choose to ignore documentation, but it cannot ignore linting errors in your CI pipeline.
As Factory AI puts it: “Linters turn human intent into machine-enforced guarantees that allow agents to plan, generate, and self-correct without waiting on humans.”
The Question I’m Wrestling With
We’re at an inflection point. In our organization, AI is writing nearly half our code. But if we don’t solve architectural drift now, we’re building a maintenance crisis for 2028.
I want to hear from this community:
What’s actually working for you?
- Are you using enhanced linting strategies that catch semantic issues?
- Have you found CI/CD automation that enforces conventions before merge?
- Are you doing context engineering (structured files that shape AI understanding)?
- Have you tried tools like Drift that learn your patterns?
What’s the right balance?
- How strict do you get without killing AI’s velocity benefits?
- Do you block PRs automatically or flag for human review?
- How do you handle the cases where AI’s “wrong” approach is actually… better?
How do you measure success?
- Convention adherence rates?
- Time saved in code review?
- Reduction in refactoring PRs?
We’re running a pilot next quarter to test AI-powered custom linting (sending diffs + style rules to an LLM for semantic review). But I’m not convinced we have the right mental model yet.
What am I missing? What’s working in your organizations?
Luis Rodriguez
Director of Engineering, Financial Services
Formerly Intel, Adobe