Skip to main content

2 posts tagged with "responsible-ai"

View all tags

The Ethics Review Gate Your AI Shipping Process Is Missing

· 9 min read
Tian Pan
Software Engineer

Most engineering teams treat ethics like they used to treat security: something you address after the feature ships, if someone complains. The parallels are uncomfortable. In 2004, SQL injection was a "we'll fix it later" problem. Today, every serious team has automated injection detection in CI. Ethics reviews in AI are at the same inflection point — and teams that don't build the gate now will learn the hard way why it exists.

The gap is not intent. It's structure. Security reviews have a 20-year head start on standardization: OWASP checklists, CVE scoring, penetration tests, mandatory sign-offs before production. Ethics reviews have none of that ceremony. Most teams have no defined trigger, no checklist, no exit criteria, and no named owner. The result: a healthcare algorithm that reduced identification of Black patients for care by over 50% not because engineers were malicious, but because no one ran disaggregated accuracy numbers before the thing went live. A recruiting model that systematically downranked resumes containing the word "women's" — trained on historical data, shipped without a fairness pass, discovered months into production. These aren't edge cases. They're what happens when ethics is a post-launch checkbox with no teeth.

The Dual Newspaper Test for AI Features: Catching the Failure Modes Your Post-Mortems Miss

· 9 min read
Tian Pan
Software Engineer

Your AI feature passed load testing. It hit the latency SLA. The rollback procedure works. Cost estimates came in under budget. Your post-mortem template has a green checkmark next to every line.

Two months after launch, the product appears in an investigative piece about discriminatory outcomes. You spend six weeks in legal review.

This is the gap the dual newspaper test is designed to close. Most engineering teams build thorough pre-ship processes for technical failures — reliability regressions, API instability, infrastructure cost blowouts. They read post-mortems about outages and optimize accordingly. But a second class of AI failures gets shipped right through those processes because it doesn't look like a bug: the feature works exactly as designed, and the harm happens anyway.