Last week, our board asked a question I wasn’t prepared to answer: “Do we actually own the copyright to our codebase?”
The question came up during M&A discussions with a potential acquirer. Their legal team wanted to know what percentage of our production code is AI-generated. I pulled our Q1 2026 telemetry: 30% of our codebase was authored by AI coding assistants, up from 22% last quarter.
Here’s the uncomfortable truth I learned: Works predominantly generated by AI, without meaningful human authorship, are not eligible for copyright protection in the United States. The US Copyright Office has been clear about this, and the Supreme Court denied certiorari on March 2, 2026, leaving that position intact.
The Scale of the Problem
This isn’t a niche issue. According to recent industry data:
- 41% of all new commercial code is AI-generated in 2026
- 76% of professional developers are using or planning to use AI coding tools
- Microsoft reports that 30% of their production code is AI-generated
- Over 40% of newly written code involves AI assistance
We’re not alone in this. But that doesn’t make the business implications any less serious.
What This Means for M&A and Valuation
During our due diligence process, the acquirer’s legal team raised three issues:
- IP Portfolio Uncertainty: If 30% of our code has no copyright protection, how do we value our technical IP?
- Licensing Risk: About 35% of AI-generated code samples contain licensing irregularities that could create legal liability
- Competitive Moats: If our “secret sauce” is mostly AI-generated using the same tools our competitors have access to, where’s the defensible IP?
They didn’t walk away from the deal, but these questions created friction we hadn’t anticipated.
Three Questions I’m Wrestling With
1. Measurement: How do we accurately track AI vs. human authorship at the code level? PR templates? Commit tags? Telemetry dashboards?
2. Governance: Should we set thresholds for AI usage in different parts of the codebase? (e.g., “core business logic must be <20% AI-generated”)
3. Documentation: What records do we need to prove “meaningful human authorship” if challenged? Code review logs? Commit histories? Developer attestations?
The Uncomfortable Trade-off
Here’s the paradox: AI coding assistants have made us 40% more productive. Our developers love them. Our velocity is up. But we’re potentially building a codebase we can’t fully claim to own.
I’m not suggesting we stop using AI tools. But I am suggesting we need governance frameworks before the next M&A conversation, not during it.
How are other CTOs thinking about this? Are you tracking AI vs. human authorship? Have you implemented policies around AI code generation? What governance frameworks are working?
Sources: