Just left the “AI for Developers” track at SF Tech Week and I have OPINIONS. Been using Cursor, Copilot, and Codeium for 6 months - let’s talk real productivity data, not marketing hype. ![]()
Sessions attended:
- GitHub Next: “The Future of AI-Assisted Development”
- Cursor team: “Building the AI-Native IDE”
- Sourcegraph panel: “Measuring Developer Productivity in the AI Age”
My Real-World Experience
Background: Full-stack engineer, Next.js/TypeScript/Python stack
Tools I’ve used:
- GitHub Copilot - 6 months
- Cursor - 3 months (current primary)
- Codeium - 2 months
- Amazon CodeWhisperer - 1 month (dropped it)
The Productivity Claims vs Reality
GitHub’s marketing: “55% faster coding”
Sourcegraph research (Oct 2025) presented at session:
Measured across 500 developers, 3 months:
- Boilerplate code: 45% faster

- CRUD operations: 35% faster

- Test writing: 40% faster

- Bug fixes: 15% faster

- Complex algorithms: 5% faster (not statistically significant)

- Architecture decisions: 0% faster (AI doesn’t help)

Aggregate: 20-25% faster coding time
BUT (the big but):
- Code review time: +15% (more code to review)
- Bug fix time: +25% (AI introduces subtle bugs)
- Net productivity: +10-15%
Not 55%. Closer to 10-15%. Still good, but not revolutionary.
Code Quality: The Uncomfortable Truth
Study from Cursor team (surprisingly honest):
Analyzed 10M lines of AI-generated code:
Bug rates:
- Human-written code: 15 bugs per 1,000 lines
- AI-suggested code (accepted as-is): 23 bugs per 1,000 lines
- AI-suggested code (edited by human): 12 bugs per 1,000 lines
The pattern:
- AI code needs human review
- Blind acceptance = more bugs
- Thoughtful editing = fewer bugs than pure human code
My experience matches this:
- I catch 60-70% of AI suggestions that would introduce bugs
- The ones I miss cause production incidents
- Example: AI generated SQL query with N+1 problem, didn’t notice until production slowdown
Security Implications
Session: “Security Risks of AI-Generated Code”
Speaker: Security researcher from Trail of Bits
Real vulnerabilities found in AI-generated code:
1. Hardcoded secrets
AI suggestion: Connection string with API key in code
Frequency: 2% of database connection code
2. SQL injection vulnerabilities
AI generates unsanitized queries
Frequency: 8% of dynamic SQL code
3. Insecure randomness
AI uses Math.random() for security tokens
Frequency: 15% of auth code
4. Missing input validation
AI trusts user input
Frequency: 25% of API endpoints
The scary part: These pass code review because they “look right” to humans who trust AI.
Remediation:
- Automated security scanning (Snyk, CodeQL)
- Security-aware AI tools
- Mandatory code review for AI-generated code
Cost: $10K/year tooling + 15% slower review process
Tool Comparison: My Experience
GitHub Copilot
- Strengths: Best for autocomplete, knows common patterns
- Weaknesses: Suggestions often generic, not context-aware
- Best for: Boilerplate, tests, simple functions
- Cost: $10/month ($20 for business)
- Rating: 7/10
Cursor
- Strengths: IDE integration, understands whole codebase, great UX
- Weaknesses: Expensive, sometimes too aggressive
- Best for: Refactoring, complex edits, codebase-wide changes
- Cost: $20/month
- Rating: 9/10 (my current choice)
Codeium
- Strengths: Free tier, fast, good autocomplete
- Weaknesses: Not as smart as Copilot/Cursor
- Best for: Budget-conscious developers
- Cost: Free (paid plans $15/month)
- Rating: 6/10
CodeWhisperer
- Strengths: AWS integration
- Weaknesses: Worst suggestions, laggy
- Best for: If you’re all-in on AWS
- Cost: Free for individual (included in AWS)
- Rating: 4/10 (dropped after 1 month)
The Cost-Benefit Analysis
Individual developer:
Costs:
- Tool: $20/month = $240/year
- Training/learning curve: 20 hours Ă— $75/hour = $1,500 one-time
Benefits:
- 10% productivity increase
- Developer writes 10,000 lines/year
- Saves: 200 hours Ă— $75/hour = $15,000/year
ROI: 6-7x (worth it for individuals)
Engineering team (50 developers):
Costs:
- Tools: $20 Ă— 50 Ă— 12 = $12,000/year
- Security tooling: $10,000/year
- Code review overhead: 15% Ă— 50 devs Ă— 10% time Ă— $150K salary = $112,500/year
- Total: $134,500/year
Benefits:
- 10% productivity across 50 devs = 5 FTE equivalent
- Value: 5 Ă— $150K = $750,000/year
ROI: 5.5x (still worth it, but diminishing returns due to overhead)
Where AI Coding Assistants Excel
From GitHub Next session - real data:
Best use cases (>40% time savings):
- Writing tests: AI understands function, generates test cases
- Boilerplate code: React components, API routes, schemas
- Documentation: AI writes docstrings from code
- Code translation: Python to TypeScript, etc.
- Regex patterns: AI is way better at regex than humans
My experience: I let AI write ALL my tests now. Saves 5-10 hours/week.
Where AI Coding Assistants Fail
Worst use cases (AI actively harmful):
1. Complex business logic
- AI doesn’t understand business requirements
- Generates plausible-looking wrong code
- Example: AI wrote tax calculation code that was subtly wrong, would have cost company thousands
2. Performance-critical code
- AI doesn’t optimize for performance
- Example: AI generated O(n²) algorithm when O(n log n) existed
3. Security-critical code
- AI copies patterns from internet (including insecure ones)
- Never trust AI for auth, crypto, payment processing
4. Architecture decisions
- AI can’t design systems
- Suggests whatever pattern it saw most in training data
- Often not appropriate for your use case
5. Code in unfamiliar languages/frameworks
- AI confidently generates wrong code
- You can’t review it effectively (don’t know the language)
- Recipe for disaster
The Learning Impact (Controversial Take)
Panel discussion: “Are AI Assistants Making Junior Devs Worse?”
Panelist (Meta engineer): “Junior engineers who learn with AI are developing differently than we did.”
Concerns:
- Less deep understanding of fundamentals
- Copy-paste mentality (used to StackOverflow, now AI)
- Difficulty debugging AI-generated code
- Weaker problem-solving skills
Counter-argument (GitHub): “Same was said about Google, StackOverflow, IDEs, high-level languages. Devs adapt.”
My take: It’s a real concern for onboarding. I spend more time teaching juniors to review AI code critically than I used to spend teaching them to write code.
Recommendation: Junior devs should write code manually for first 6-12 months before using AI assistants.
Organizational Adoption Challenges
From Sourcegraph session:
% of developers at companies with AI tools who actually use them daily:
- Individuals (self-paid): 85%
- Companies <50 devs: 60%
- Companies 50-500 devs: 35%
- Companies 500+ devs: 20%
Why low adoption at large companies?
- Security concerns: Legal won’t approve (code leaves environment)
- IP concerns: Who owns AI-generated code?
- Compliance: Regulated industries prohibit external AI tools
- License issues: AI trained on GPL code, copyright unclear
- Resistance to change: Senior devs skeptical
Successful adoption pattern:
- Start with security-approved pilot (10-20 devs)
- Measure productivity (real metrics, not vibes)
- Address security/IP with legal
- Roll out gradually with training
The IP and Licensing Minefield
GitHub Legal session (scary):
Open questions (no legal consensus yet):
-
Who owns AI-generated code?
- Developer?
- Company?
- AI vendor?
- Unknown
-
Copyright infringement risk
- AI trained on copyrighted code
- Output may reproduce copyrighted code
- Liability unclear
-
License contamination
- AI suggests GPL code in your proprietary project
- Is your code now GPL?
- Courts haven’t decided
GitHub’s position: “GitHub Copilot doesn’t infringe. We provide legal coverage.”
Reality: No court cases yet. First lawsuit will set precedent.
Risk-averse companies: Ban AI tools until legal clarity
My Current Workflow
What I use AI for:
- Writing tests (AI writes, I review): 80% time savings
- Boilerplate (components, routes, schemas): 60% time savings
- Documentation (AI writes docstrings): 70% time savings
- Explaining unfamiliar code: Huge value, hard to quantify
What I don’t use AI for:
- Business logic (too risky)
- Security code (never)
- Architecture (AI can’t)
- Complex algorithms (AI gets it wrong)
My rule: If I can’t verify the code in 30 seconds, I don’t accept the AI suggestion.
The Future: Predictions from GitHub Next
What’s coming (next 12-18 months):
-
Codebase-aware AI (already here with Cursor, getting better)
- Understands your entire project
- Suggests changes across files
- Maintains consistency
-
Test generation (early stage)
- AI writes comprehensive test suites
- Identifies edge cases
- Generates test data
-
Bug detection (experimental)
- AI reviews PRs for bugs
- Security vulnerability detection
- Performance issues
-
Autonomous coding agents (research phase)
- Give AI a task, it writes code
- Self-corrects based on test failures
- Human reviews final result
Timeline to production: 2-5 years for most of these
My Recommendations
For individual developers:
Use AI coding assistants (ROI is clear)
Start with Cursor or Copilot
Use for tests, boilerplate, docs
Don’t blindly accept suggestions
Don’t use for security-critical code
For engineering teams:
Run pilot program (10-20 devs, 3 months)
Measure actual productivity (not perception)
Address security/legal concerns upfront
Train developers on effective AI use
Don’t force adoption (let devs choose)
Don’t skip code review for AI code
Bottom line: AI coding assistants are real productivity boost (10-15%), not revolutionary (not 10x), and require thoughtful adoption.
Worth $20/month? Absolutely.
Going to replace developers? Not even close.
Who else is using AI coding tools? What’s your experience?
David ![]()
SF Tech Week - “AI for Developers” track, Moscone Center
Note: Sharing developer perspective from our engineering team
Sources:
- Sourcegraph “Developer Productivity Study” (Oct 2025)
- GitHub Next “Future of Coding” (Oct 2025)
- Trail of Bits “AI Code Security” research