Mitigating Bias in AI Hiring Tools

NEWSLETTER VOLUME 2.46

|

November 26, 2024

Editor's Note

Mitigating Bias in AI Hiring Tools

 

A University of Washington study that tested AI resume screening tools found the tools discriminate—like the resumes of Black men were never selected discriminate. The article below from Fisher Phillips is a great discussion of the study and how employers can approach the issue. The only thing that I disagree with is that the results are shocking.

 

How do we end up with hiring tools that systematically exclude all Black men and most women (of any hue)? It's the data. And the data is based on human actions and decisions, which are biased because they're human. (Not yours, of course.) Humans like things that are familiar and feel anxious about things that are not familiar. So the people doing the deciding tend to pick people that are more familiar to them and not pick people who aren't. All humans do this. It even has a name: familiarity bias.

 

Discrimination in the legal sense comes in when the deciders tend to be one particular race and/or gender so that race and gender ends up weighting the results. If you strip away everything else, it usually comes down to who is the decider.

 

In diversity discussions, we often focus on the people who are excluded in processes like hiring and promotion. This makes sense because that's where the potential risk is for the employer and where the harm is for employees.

 

But in tech discussions, it's also useful to focus on what is making the data that way because it may open up new approaches to mitigating bias in the data our hiring tech uses.

I don't know enough about the logistics of cleaning data and addressing bias in datasets to say anything useful about the process itself. I do know that trying to change a biased data set is not like pulling a couple ugly cherries out of the pie before you slice and serve it. It's more like trying to remove the shortening from the crust.

 

Yet, one set of data that seems to be missing from much of people analytics on diversity and hiring is demographic data on recruiters and hiring managers. If we want to clearly see hiring outcomes with tech and analytics, we can't just clean the data to mitigate bias, we also need to expand the data to cover the full system. It's not just the data; it's the deciders.

 

- Heather Bussing


New Study Shows AI Resume Screeners Prefer White Male Candidates: Your 5-Step Blueprint to Prevent AI Discrimination in Hiring

by Amanda Blair and Karen Odash

at Fisher Phillips

 

A new study reveals that popular AI-based resume screening tools often favor White and male candidates, showing that resumes tied to White-associated names were preferred 85% of the time. The University of Washington Information School study found that AI resume screening may inadvertently replicate – or even amplify – existing biases against certain demographics, and the implications are clear: without targeted safeguards, AI tools may lead to discriminatory hiring practices, even if inadvertent. This Insight breaks down the study’s findings, explains why they matter for HR professionals, in-house counsel, and employers, and provides five practical steps you can take to avoid AI bias at your workplace.

Shocking Results From AI Hiring Study

The University of Washington study, presented at the October Association for the Advancement of Artificial Intelligence/Association for Computer Machinery (AAAI/ACM) Conference on AI, Ethics, and Society, highlights the biases present in several leading open-source AI resume-screening models.

Researchers Kyra Wilson and Aylin Caliskan audited three popular large language models (systems trained on a general corpus of language data rather than data specific to hiring tasks) using 554 resumes and 571 job descriptions, with over 3 million combinations analyzed across different names and roles. They then changed up the resumes, swapping in 120 first names generally associated with individuals who are male (such as John, Joe, Daniel), female (Katie, Rebecca, Kristine, etc.), Black (Kenya, Latrice, Demetrius, etc.), or White (Hope, Stacy, Spencer, etc.). They then submitted the resumes for a series of positions ranging from chief executive to sales workers.

The results were stark:

  • Resumes with White-associated names were selected 85% of the time for the next hiring step, while resumes with Black-associated names were only preferred 9% of the time.
  • Resumes with male-associated names were preferred 52% of the time, even for roles with a traditionally high representation of women – like HR positions (77% women) and secondary school teachers (57% women).
  • Resumes with White female names were chosen over those with Black female names, by a margin of 48% to 26%.
  • Black men faced the greatest disadvantage, with their resumes being overlooked 100% of the time in favor of other candidates.

“Garbage In, Garbage Out”

Researchers attribute these biased outcomes to the data used to train the AI models. AI systems inherently mirror the patterns present in their training datasets. When these datasets are drawn from sources with historical or societal inequities, the AI system is likely to replicate or even amplify those inequities, leading to biased decision-making. For instance, if past hiring data or generalized language data is skewed to favor White or male candidates due to longstanding social privileges, the AI model will absorb and perpetuate this bias in its recommendations, leading to unfair preferences for candidates from those groups.

Moreover, because the underlying training data often incorporates a wide array of data from sources that may include biased language or assumptions, AI models trained on such data inherit biases that may not be immediately visible but become evident in their outputs. This highlights the vulnerability of AI use in applications, like hiring, where fairness is critical.

When we feed in “garbage” – data influenced by historic inequalities – we get “garbage” out, meaning decisions that are inherently flawed or biased. Consequently, AI tools that lack sufficient attention to diversity and equity in their training data risk becoming automated gatekeepers of discrimination, systematically disadvantaging qualified candidates from underrepresented groups.

Ultimately, this “garbage in, garbage out” effect underscores the need for companies to scrutinize both the data used in AI systems and the models themselves. Without interventions to correct these biases, companies risk reinforcing the very inequalities that AI was initially meant to help eliminate.

Here’s Why Employers Should Care

The stakes are high for employers adopting AI when hiring. While the technology promises efficiency, the potential for biased outcomes carries significant legal, ethical, and reputational risks.

  • Discriminatory hiring practices can lead to costly lawsuits or government investigations into your workplace practices.
  • The EEOC and other agencies have already said that existing laws cover so-called AI discrimination, so you shouldn’t stick your head in the sand waiting for new protections to be enshrined into the law before acting.
  • Even unintentional discrimination can lead to an adverse finding, so you won’t be able to fully defend your organization by saying you didn’t mean to discriminate.
  • The government has also already indicated that employers can’t simply point the finger at their AI vendors and hope to deflect responsibility. If your technology ends up discriminating against applicants or workers, agencies have said that your company could be on the hook.
  • By screening out candidates who don’t fall into the White male category, you will lose out on an array of diverse, qualified talent that can strengthen your organization.
  • The negative publicity that could result from accusations of AI bias could be costly. Your recruitment and retention efforts could take a real hit, as could your reputation among clients and customers, if you are outed as an employer that uses AI technology that discriminates against applicants.

5 Best Practices Checklist to Avoid AI Discrimination in Hiring

As the study notes, simply removing names from resumes may not address the issue. Resumes contain various elements, such as educational institutions, locations, group memberships, and even language choices (e.g., women often use terms like “cared” or “volunteered,” while men might use “repaired” or “competed”) that can indicate an individual’s race or gender, potentially perpetuating biased outcomes.

The good news is that there are some simple steps you can take to minimize the chance of AI bias at your workplace. Here are five best practices you should consider implementing at your organization.

1. Regularly Audit AI Models for Bias

Conduct regular audits of AI tools to detect racial, gender, and intersectional biases. Collaborate with third-party evaluators or leverage in-house resources to regularly examine screening outcomes and ensure fairness in the hiring process.

2. Consider Data Transparency and Fairness

Ensure that the data used to train your models is balanced and representative, and prioritize models with built-in transparency. Ask vendors the right questions so they can demonstrate their clear efforts to minimize biases in training datasets.

3. Avoid Over-Reliance on Automated Decisions

“Automation bias” is a real cognitive bias which can lead people to trust AI decisions more than human judgment. This can reinforce biased outcomes when hiring, as your managers may rely heavily on AI-driven recommendations, believing them to be “better” than anything they could do on their own. Address these risks head on. Make sure to integrate – and mandate – human oversight into AI decisions. Train your hiring managers to use AI tools as supplementary rather than primary decision-makers.

4. Adopt Inclusive Job Descriptions and Screening Criteria

Ensure your job descriptions and selection criteria are neutral and inclusive. Remove unnecessary criteria that may unfairly disadvantage certain candidates, such as rigid experience or narrowly defined skills that may unfairly disadvantage certain applicants. Remember, if you input these unintentional biases into your AI systems, they will spit biased results right back at you.

5. Implement Data-Driven Safeguards

Regularly analyze the outcomes of your AI screening tools, comparing demographic results to identify and address any biases. Use this data to make necessary adjustments to screening algorithms or criteria.

It's Easy to Get Started

Transform compensation at your organization and get pay right — see how with a personalized demo.