The Culture Fit Problem: How Hiring Language Encodes Bias and What to Replace It With

The Problem With “Culture Fit”

Every engineering leader I know has sat in a debrief where someone said “they just don’t seem like a good culture fit” — and then the room nodded and moved on. I’ve been in those rooms. I’ve nodded. And I’ve spent the last several years trying to understand why that phrase is one of the most dangerous things you can say in a hiring process.

Let’s start with what “culture fit” was supposed to mean.

The Original Intent

When organizational psychologists first introduced culture fit as a hiring criterion, the idea was legitimate: does this person share the company’s core values? Do they approach work in ways that are compatible with the team’s working style? Can they thrive in this environment? These are real questions. A company that values deep asynchronous thinking will genuinely struggle to integrate someone who needs constant real-time collaboration. A startup that moves at breakneck speed will exhaust someone who does their best work through deliberate, methodical process.

Used correctly, culture fit was about compatibility — not conformity.

What It Became

Here’s what research has consistently shown: in practice, “culture fit” is a proxy for demographic similarity.

The MIT and University of Chicago audit studies are the most damning. In callback-rate studies where identical resumes were submitted with stereotypically Black-sounding vs. white-sounding names, callback rates differed by 50% or more. But the effect is not just about race — it’s about every dimension of perceived similarity. Interviewers rate candidates higher on “culture fit” when they share hobbies, attended similar schools, have similar communication styles, and come from similar socioeconomic backgrounds.

A 2012 study by Lauren Rivera (“Hiring as Cultural Matching”) documented elite professional services firms explicitly using culture fit to select candidates who shared interviewers’ leisure activities — squash, sailing, certain kinds of travel. The work itself was barely discussed.

This is what culture fit screening has become: an unstructured, unaudited, bias-amplifying mechanism that lets interviewers act on gut feeling while believing they’re making principled decisions.

The Business Irony

Here’s what makes this particularly frustrating from a strategic standpoint: the business case for cognitive diversity is well-established. Teams with diverse perspectives identify more potential failure modes, generate more creative solutions, and make fewer correlated errors. McKinsey’s diversity research consistently shows that companies in the top quartile for ethnic and gender diversity outperform on profitability.

So we have a hiring criterion — culture fit — that was designed to build cohesive teams, but instead screens out the cognitive diversity that makes teams actually perform better. We’re optimizing for comfort and producing underperformance.

Culture Add: A Better Frame

“Culture add” does not mean abandoning the goal of cultural cohesion. It means asking a different question: not “does this person fit what we already are” but “does this person share our core values AND bring something we’re missing?”

This requires two things that culture fit never required: (1) being explicit about what your core values actually are, and (2) being honest about what gaps exist in your current team.

If your core values are things like “ships high-quality work,” “gives and receives feedback directly,” and “takes ownership through ambiguity” — those are evaluable criteria. You can build interview questions around them. You can score candidates against them. You can audit your pass rates across demographic groups.

Structural Changes That Actually Work

Here are the practices that have evidence behind them:

Structured interviews with shared rubrics. Every interviewer scores the same dimensions using the same criteria defined before the interview. This dramatically reduces the surface area for culture-fit rationalization.

Diverse interview panels. Not tokenism — genuine representation across demographic dimensions, seniority, and working style. Research shows diverse panels produce more consistent, less biased evaluations.

Work-sample tests. A take-home problem, a pair-programming session, or a system design exercise evaluated against explicit criteria tells you far more about job performance than how comfortable you feel in a 45-minute conversation.

Blind resume screening. Remove names, graduation years, and school names from initial screening. Several studies show this alone significantly increases demographic diversity in interview pools.

Calibration sessions. After each hiring cycle, review your pass rates by demographic group. If you cannot see disparities, you cannot address them.

What Cohesion Without Conformity Looks Like

The best engineering organizations I’ve seen achieve something harder than either culture fit OR culture add in isolation: they have very explicit, well-understood values that people are genuinely selected for — and they hire people who are genuinely different from each other within that values framework.

That requires doing the harder work of defining what you actually care about, building evaluation systems that measure it, and having the discipline to override gut feeling when the process contradicts it.

It’s more work than nodding in a debrief. It’s also how you build a team that’s actually capable of the work ahead.

This post hits really close to home for me. I’ve been on the receiving end of the “culture fit” rejection and it stung in a way I couldn’t fully articulate at the time. :pensive_face:

I was interviewing at a company a few years back — cleared every technical round, got genuinely positive feedback on every dimension they could actually measure, and then in the final loop one interviewer said I seemed “a bit intense” and another said they weren’t sure I’d “vibe” with the team. I got the rejection email two days later citing culture fit. I’m a woman who speaks directly and pushes back on bad ideas in real time. Make of that what you will.

When I became a hiring manager I was determined to do better, and here’s what I actually changed:

No open-ended culture fit discussions in debriefs. If someone raises it, I ask them to translate it into a specific observed behavior from the interview. “They seemed intense” becomes a question: intense in what way, and is that actually a problem for this role?

I defined our values explicitly before we started hiring. Not platitudes — actual behavioral indicators. “Communicates directly” means: gave specific feedback when asked, didn’t hedge criticism of a design choice.

I started noticing when everyone on the panel looked the same. When your panel is homogeneous, your “culture fit” signal is a mirror test. :roll_eyes:

It’s not a perfect process and I still catch myself having gut reactions I have to interrogate. But at least now there’s a process to interrogate them against. The culture fit vibe check has nowhere to hide when you’ve written down what you’re actually looking for. :flexed_biceps:

The research picture on this is pretty grim, and the tricky part is that measuring bias in your own hiring process is genuinely hard.

The audit studies Keisha cites are the gold standard precisely because they use identical resumes — the only variable is the name. In those controlled conditions the signal is unambiguous. But inside your own company you don’t have a controlled experiment. You have a noisy process with dozens of confounding variables, and the people who would need to audit it are often the same people whose judgment is being audited.

A few things that actually help:

Track pass rates by demographic group at each stage. You need enough volume for this to be meaningful, but even small samples can reveal dramatic drop-off patterns. If candidates from a particular group are advancing through technical screens at the same rate but failing culture/values interviews at higher rates, that’s a signal worth investigating.

A/B test your job descriptions. Tools like Textio and Gender Decoder will show you language that systematically discourages applications from underrepresented groups before you’ve even started the process. The bias starts with who applies.

Structured interviews improve consistency even without demographic data. Studies consistently show structured interviews predict job performance better than unstructured ones. The culture-fit reduction is partly a side effect — when you’re scoring against criteria, there’s less room for vibes.

The honest answer is that most companies don’t have enough hiring volume or demographic diversity in their applicant pools to run properly powered studies. But you can still build process controls that make bias harder to act on, even if you can’t measure it perfectly.

About two years ago I replaced “culture fit” entirely in my team’s process and replaced it with five explicit values-based criteria. Each criterion has two behavioral interview questions and a scoring rubric. “Culture fit” literally does not appear in any of our interview documentation.

The pushback I got from the team was more interesting than I expected. The resistance was not from people who wanted to discriminate — it was from people who genuinely believed their gut instinct was a signal worth preserving. “I’ve been doing this for ten years, I know when someone’s going to be a good teammate” is a real belief held by experienced engineers who are not bad people.

My answer to that: your gut is data, but it’s not sufficient. Translate it. If your gut says someone will be a bad teammate, I want to know specifically what you observed that generates that prediction. Is it that they interrupted you repeatedly? That they couldn’t acknowledge any tradeoff in their technical approach? That they had no questions about the team? All of those are translatable into evaluable behaviors. “I just didn’t feel it” is not.

After two years, a few observations: Our time-to-fill went up slightly (more process takes more time). Our offer acceptance rate went up more than time-to-fill went up. And the last three engineers we hired who came from backgrounds different from our team’s demographic norm have all been promoted or given expanded scope within 18 months.

Correlation, not causation. But I’ll take the correlation.