I just finished integrating Mintlify into our developer tools platform at TechFlow, and I’m genuinely conflicted about what I’m seeing. On one hand, we’re saving roughly 15 hours per week on documentation maintenance. On the other hand, I’m realizing there’s a really specific line between what AI does brilliantly and where it completely falls apart.
The Promise
The pitch was compelling: point an AI tool at your codebase, and it automatically generates documentation that stays in sync with your code. No more outdated function signatures. No more parameter descriptions that reference arguments that don’t exist anymore. The documentation lives and breathes with your code.
The Reality (It’s Complicated)
Here’s what I’ve learned over the past six weeks:
Where AI excels: API reference documentation. Mintlify absolutely nails this. It reads our TypeScript types, extracts JSDoc comments, and produces clean, accurate API references. It caught three instances where our manual docs referenced old parameter names that we’d refactored months ago. That alone justified the investment.
Where AI struggles: Conceptual guides and architectural documentation. We needed to document a migration path from our v2 to v3 API. The AI-generated version was technically accurate but completely missed the why behind the changes. It listed what changed but didn’t explain the business reasoning, the architectural decision-making process, or the edge cases that drove our design.
The Pattern I’m Seeing
AI handles the “what” brilliantly:
- Function signatures
- Parameter types
- Return values
- Basic usage examples
But humans are still essential for the “why” and “how”:
- Architectural decisions and tradeoffs
- Migration guides and upgrade paths
- Edge cases and gotchas
- Conceptual explanations that tie features together
What This Means for Our Workflow
We’ve settled into a rhythm where AI is our first draft generator and consistency checker. For API references, we basically just review and publish. For guides and tutorials, we use AI output as a starting point, then heavily edit to add context, examples from real usage, and the kind of judgment that only comes from actually building with the tools.
I’m not writing as much documentation from scratch anymore. Instead, I’m editing, curating, and adding the human layer that makes docs actually useful rather than just accurate.
The Question I’m Wrestling With
This feels like a fundamental shift in what “writing documentation” means. We’re moving from author to editor. From creator to curator. And I think that’s probably fine? Maybe even better? But I’m curious how others are thinking about this balance.
How are you handling AI-generated documentation? What’s your threshold for “good enough” versus “needs human refinement”? And are there types of docs you’d never trust to AI alone?