I need to share something that’s been keeping me up at night, and I’m hoping this community can help us figure out next steps.
My team built an AI-powered resume screening tool to help our recruiting team handle volume more efficiently. The pitch was compelling: remove human bias from initial resume review by using machine learning to identify qualified candidates based on objective criteria. We trained the model on our historical hiring data – specifically, resumes of people who were hired and went on to be successful employees.
Last month, we ran a bias audit before wider deployment. Standard practice, right? Check for disparate impact across demographic groups. What we found made me sick.
The model was systematically downranking candidates from women’s colleges and HBCUs. Not by a little – by a lot. A resume from Smith College or Spelman was scored 30-40% lower than an identical resume from a comparable co-ed or predominantly white institution. We’re talking about talented engineers who went on to work at top companies, but our “objective” AI said they weren’t qualified.
Here’s the technical explanation: our training data reflected historical hiring patterns. For the past decade, our company (like most tech companies) hired heavily from a short list of tier-1 CS programs – Stanford, MIT, Carnegie Mellon, Berkeley. Very few hires came from women’s colleges or HBCUs, not because candidates weren’t qualified, but because we weren’t recruiting there. The AI learned that “successful candidate” equals “traditional tech pipeline school.”
So our tool marketed as “removing bias” was actually encoding and amplifying systemic bias. It learned our historical blind spots and turned them into algorithmic gatekeeping.
We immediately paused deployment. We’re now working with bias detection specialists to rebuild the training approach. But I keep thinking about the bigger picture: how many companies have deployed similar tools without auditing them? How many qualified candidates are being filtered out by algorithms that claim to be objective?
The whole premise feels flawed now. We wanted AI to fix human bias, but we trained it on data created by biased humans. Garbage in, garbage out – except the garbage is people’s careers.
Here’s what I’m wrestling with:
Is truly unbiased AI hiring even possible? Or are we just creating more sophisticated ways to justify the same patterns?
What should audit processes look like for hiring algorithms? We only caught this because we ran demographic analysis, which many companies skip.
Should companies be required to disclose when they use AI in hiring decisions? Candidates have no idea they’re being filtered by algorithms.
What’s the alternative? Going back to fully manual review has its own bias problems, and we don’t have the capacity to read every resume.
I’d love to hear from others who’ve worked with hiring algorithms – what red flags should we be watching for? What’s your experience been with trying to make these tools fair?
This feels like one of those moments where the tech industry’s faith in technological solutions to social problems gets exposed as naive. But I also don’t want to give up on the idea that we can build better tools. I just don’t know what “better” looks like anymore.