Saturday, June 29, 2024
- Advertisement -
Saturday, June 29, 2024

Disability bias in ChatGPT-based resume screening

AI systems may harbour inherent biases that can disadvantage candidates with disabilities

Must Read

- Advertisement -
  • UW research revealed ChatGPT and other AI tools consistently rank resumes with disability-related honours and credentials lower than their counterparts without such information.
  • As these technologies become more pervasive, they must be designed and implemented with a keen awareness of their impact on marginalised groups, including individuals with disabilities.

In today’s increasingly technology-driven job market, the use of artificial intelligence (AI) tools in the hiring process has become a common practice.

However, as the University of Washington (UW) research has revealed, these AI systems may harbour inherent biases that can disadvantage candidates with disabilities.

Addressing this issue is not only a moral imperative but also a necessary step toward creating a more inclusive and equitable job market.

The study conducted by UW researchers, led by doctoral student Kate Glazko at the UW’s Paul G. Allen School of Computer Science & Engineering, shed light on the concerning reality that ChatGPT and other AI tools consistently ranked resumes with disability-related honours and credentials lower than their counterparts without such information.

When prompted to explain its reasoning, the system revealed biased perceptions and harmful stereotypes regarding individuals with disabilities.

“Ranking resumes with AI is starting to proliferate, yet there’s not much research behind whether it’s safe and effective,” she said.

However, she said that some of GPT’s descriptions would colour a person’s entire resume based on their disability and claimed that involvement with diversity, equity and inclusion (DEI) or disability is potentially taking away from other parts of the resume.

 “In a fair world, the enhanced resume should be ranked first every time,” said senior author Jennifer Mankoff, a UW professor in the Allen School.

“I can’t think of a job where somebody who’s been recognised for their leadership skills, for example, shouldn’t be ranked ahead of someone with the same background who hasn’t.”

These findings underscore the urgent need for a deeper examination of the potential pitfalls of AI-powered resume screening. As these technologies become more pervasive, they must be designed and implemented with a keen awareness of their impact on marginalised groups, including individuals with disabilities.

Researchers, cognizant of the inherent biases present in large language models like GPT-4, have sought to explore strategies for cultivating more inclusive and equitable AI systems.

The study described in the provided text highlights a commendable effort to address ableist biases within the GPT-4 model. By leveraging the GPTs Editor tool, the researchers were able to customise the chatbot’s responses to align with disability justice and DEI principles.

The approach, which did not require complex coding, demonstrates the potential for accessible and impactful interventions in the field of AI development.

The experimental results are encouraging, as the enhanced chatbot ranked the modified CVs higher than the control CV in a majority of the trials.

 “People need to be aware of the system’s biases when using AI for these real-world tasks. Otherwise, a recruiter using ChatGPT can’t make these corrections, or be aware that, even with instructions, bias can persist,” Glazko said.

Encouragingly, the UW researchers’ efforts to customise the AI tool with written instructions directing it to avoid ableist biases yielded promising results. While the system was not able to eliminate the bias for all the disabilities tested, it did show a significant improvement in ranking resumes with five out of the six implied disabilities higher than before.

“It is so important that we study and document these biases,” Mankoff said.

“We’ve learned a lot from and will hopefully contribute back to a larger conversation — not only regarding disability but also other minoritised identities — around making sure technology is implemented and deployed in ways that are equitable and fair.”

Related Posts:


Discover more from TechChannel News

Subscribe to get the latest posts to your email.

Advertisement
- Advertisement -

Latest News

Freshworks unveils UAE data centre to focus on clients in Middle East and Africa

Data residency has always been a topic of conversation with customers in the region

Five disturbing security threats lurking in TikTok comments

What steps will you take to safeguard your TikTok interactions?

Russian hacker group gains access to more Microsoft customer emails

Microsoft is set to notify its additional customers whose emails were compromised by Midnight Blizzard
- Advertisement -

More Articles

- Advertisement -
×