arXiv Open Access 2025

Evaluating Bias in LLMs for Job-Resume Matching: Gender, Race, and Education

Hayate Iso Pouya Pezeshkpour Nikita Bhutani Estevam Hruschka
Lihat Sumber

Abstrak

Large Language Models (LLMs) offer the potential to automate hiring by matching job descriptions with candidate resumes, streamlining recruitment processes, and reducing operational costs. However, biases inherent in these models may lead to unfair hiring practices, reinforcing societal prejudices and undermining workplace diversity. This study examines the performance and fairness of LLMs in job-resume matching tasks within the English language and U.S. context. It evaluates how factors such as gender, race, and educational background influence model decisions, providing critical insights into the fairness and reliability of LLMs in HR applications. Our findings indicate that while recent models have reduced biases related to explicit attributes like gender and race, implicit biases concerning educational background remain significant. These results highlight the need for ongoing evaluation and the development of advanced bias mitigation strategies to ensure equitable hiring practices when using LLMs in industry settings.

Topik & Kata Kunci

Penulis (4)

H

Hayate Iso

P

Pouya Pezeshkpour

N

Nikita Bhutani

E

Estevam Hruschka

Format Sitasi

Iso, H., Pezeshkpour, P., Bhutani, N., Hruschka, E. (2025). Evaluating Bias in LLMs for Job-Resume Matching: Gender, Race, and Education. https://arxiv.org/abs/2503.19182

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓