1. Introduction and Research Questions
AI in recruitment raises concerns about algorithmic bias and employment discrimination. AI systems, while designed to improve efficiency, can perpetuate existing biases from historical data, leading to discriminatory outcomes for marginalized groups. Research shows AI algorithms can reflect racial, gender, and socioeconomic biases, amplifying societal inequalities. For example, Amazon’s AI recruitment tool favored male candidates not female, highlighting the potential for bias in AI systems[1][2]. These biases have broader societal implications, contributing to wage gaps and career limitations. Ethical strategies are needed to mitigate bias, ensuring AI promotes inclusivity rather than discrimination.
This research employs a systematic literature review to explore AI and employment discrimination, focusing on algorithmic bias in HR practices. By analyzing existing literature and case studies, the study aims to understand the nature of bias, its impact on hiring, and potential measures to address these challenges. As AI increasingly influences recruitment, critical research questions emerge regarding the mechanisms of bias and its broader implications. This investigation seeks to uncover the complexities of AI-related employment discrimination, contributing to the ongoing discourse on fairness in AI and its potential consequences for the future of work.
1.1. What are the underlying causes of algorithmic bias in AI recruitment systems, and how can we disentangle the intricate web of factors contributing to this bias?
Understanding the root causes of algorithmic bias involves examining training datasets, algorithm design principles, and human decisions in AI implementation. Biases can arise from historical hiring practices in data or subjective choices during model development. Identifying these causes is essential for addressing the issue and forms the critical first step toward developing effective solutions to mitigate bias in AI systems.
1.2. How does algorithmic bias affect hiring outcomes across different demographic groups?
It is crucial to analyze how biased algorithms influence hiring practices among various demographic cohorts. Evidence suggests that marginalized groups—including women and racial minorities—are frequently disadvantaged by AI-driven recruitment processes. By investigating specific case studies, such as Amazon's recruitment tool which exhibited gender bias, researchers can quantify disparities in hiring outcomes and evaluate their broader implications for workforce diversity.
1.3. What measures can be implemented to mitigate algorithmic bias in AI hiring practices?
Identifying potential solutions to reduce bias is crucial for promoting fairness in AI recruitment. This question prompts an exploration of existing strategies, such as data anonymization, algorithmic auditing, and the incorporation of fairness-aware machine learning techniques. Assessing the effectiveness of these measures will contribute to the development of best practices for organizations that utilize AI in hiring.
1.4. What role do regulatory frameworks play in addressing algorithmic bias in employment discrimination?
Investigating the impact of existing laws and regulations surrounding AI and employment discrimination can provide insights into how policies can evolve to better protect candidates from biased hiring practices. This question encourages an exploration of legal precedents, proposed legislation, and ethical guidelines that govern the use of AI in recruitment.
2. Literature Review
The integration of AI in Human Resource Management (HRM) has introduced a new paradigm in recruitment practices. AI-driven recruitment tools are designed to streamline the hiring process, reduce costs, and minimize human biases. However, these tools are not without their flaws, with algorithmic bias emerging as a significant concern. This literature review examines the current state of AI in recruitment, the sources of algorithmic bias, and the potential for discrimination.
The adoption of AI in recruitment has been lauded for its potential to objectify the selection process. Fabris et al. suggest that AI can analyze vast amounts of data to identify the best candidates, thus reducing the cognitive biases inherent in human decision-making[3]. However, this optimism is tempered by the reality that AI systems are trained on historical data, which may contain biases that are then amplified by the algorithms [4]. Algorithmic bias can arise from various sources, including the data used to train the AI, the features selected by engineers, and the inherent biases of the designers. Adams-Prassl et al. argue that directly discriminatory algorithms are those that are explicitly programmed to discriminate, while indirectly discriminatory algorithms learn to discriminate from biased data[5]. Ali et al. demonstrate how optimization algorithms can lead to discriminatory outcomes, such as Facebook's ad delivery system favoring certain groups over others. One of the most concerning aspects of algorithmic bias is its impact on gender, race, and personality[6]. Ajunwa highlights the paradox of automation as an anti-bias intervention, where AI systems may inadvertently perpetuate gender stereotypes. Ameri et al. conducted a field experiment revealing disparities in employer hiring behavior towards individuals with disabilities, which could be exacerbated by AI if not carefully designed.
AI integration in HRM has revolutionized recruitment by streamlining processes, reducing costs, and minimizing human biases. However, algorithmic bias remains a significant issue. This review explores the current use of AI in recruitment, the sources of bias, and the potential for discrimination. The digital economy has shifted statistical discrimination theory from traditional to intelligent hiring, using historical data to predict future outcomes.
Algorithmic bias refers to the systematic and replicable errors in computer systems that lead to unequal discrimination based on legally protected characteristics, such as race and gender [7]. When assessments consistently overestimate or underestimate a particular group’s scores, they produce “predictive bias”[8]. Unfortunately, these discriminatory results are often overlooked or disregarded due to the misconception that AI processes are inherently “objective” and “neutral”
Within the recruitment process, algorithmic bias can manifest concerning gender, race, color, and personality. Gender stereotypes have infiltrated the “lexical embedding framework” utilized in natural language processing (NLP) techniques and machine learning (ML). Munson’s research indicates that “occupational picture search outcomes slightly exaggerate gender stereotypes, portraying minority-gender occupations as less professional”[9].
The impact of gender stereotypes on AI hiring poses genuine risks [10]. In 2014, Amazon developed an ML-based hiring tool, but it exhibited gender bias. The system did not classify candidates neutrally for gender [11]. The bias stemmed from training the AI system on predominantly male employees’ CVs [10]. Accordingly, the recruitment algorithm perceived this biased model as indicative of success, resulting in discrimination against female applicants [12]. The algorithm even downgraded applicants with keywords such as “female” [13]. These findings compelled Amazon to withdraw the tool and develop a new unbiased algorithm. However, this discrimination was inadvertent, revealing the flaws inherent in algorithmic bias that perpetuates existing gender inequalities and social biases[14].
To address algorithmic bias, a multifaceted approach is required, combining technical solutions like fair datasets, algorithmic transparency, and bias-detecting tools with managerial strategies. These include internal ethics governance, diversity in AI teams, and external oversight via third-party audits and regulations. Fabris et al. advocate for a multidisciplinary approach involving technical, legal, and ethical frameworks to mitigate bias [15].
3. Discussion
The selected literature was analyzed thematically, with a focus on the sources, types, and potential solutions for algorithmic bias in AI recruitment. The analysis aimed to identify common themes and trends, as well as gaps in the current research.
3.1. Various contributing factors in algorithmic bias in AI recruitment need requires examining
The data sources reflect a hidden history of bias. The datasets used to train AI algorithms play a crucial role in shaping their behavior. Historical hiring data often reflects the biases and societal norms of the past, including under-representation of certain groups, gender and racial stereotypes, and historical patterns of discrimination. When AI systems learn from this biased data, they are likely to perpetuate and amplify these existing inequalities. If the training data primarily consists of male-dominated fields, the algorithm may prioritize male candidates for similar positions in the future. This can perpetuate gender disparities in the workforce and limit opportunities for qualified women.
3.2. Algorithm design choices by AI developers significantly influence the potential for bias
Decisions about candidate evaluation features, model architecture, and evaluation metrics can introduce bias in AI-driven hiring. For example, if an algorithm uses features linked to certain demographics, it may unintentionally disadvantage other groups. Similarly, the choice of evaluation metrics can prioritize certain traits, leading to biased outcomes.
While AI systems are often perceived as objective and neutral, they are ultimately created by humans who may hold unconscious biases. The role of unconscious bias comes from human influence. Data selection, feature prioritization, and even the development process's language can unintentionally encode these biases into the algorithms. Developers may inadvertently prioritize certain keywords or phrases that are more commonly associated with certain demographic groups, leading to biased candidate selection. Quantifying disparities is a key step. Research indicates that women and racial minorities are often at a disadvantage when facing AI-driven recruitment processes.
MIT researcher Joy Buolamwini exposed significant racial biases in facial recognition technology. Her study found that commercial systems achieved 99% accuracy for white men but had higher error rates for individuals with darker skin tones. Around 35% of black women, including Oprah Winfrey and Michelle Obama, were misidentified. The ACLU's report on Amazon's Rekognition system also revealed misidentifications, including erroneous matches between Congress members and criminal suspects.
3.3. Case studies of algorithmic bias in hiring reveal real-world consequences of biased algorithms
The Amazon recruitment tool case, for example, highlighted how historical data and subjective design choices could lead to gender discrimination. Amazon has been developing a system since 2014 to ease the burden of processing tens of thousands of resumes during its continuous growth. The design team used 500 samples to improve their recruitment system, but certain keywords negatively impacted applicant rankings. The system unintentionally perpetuated human biases through deep learning. Despite efforts to adjust, the outcomes remained uncertain. This highlights how biases in traditional recruitment can be ingrained and learned by AI. According to Reuters, male employees at Amazon, Facebook, Apple, Google, and Microsoft make up 60%, 64%, 68%, 69%, and 74%, respectively, with even higher male representation in technical roles, especially at Microsoft (81%).
Algorithmic bias in hiring not only affects individual outcomes but also perpetuates societal issues like wage gaps, limited career advancement, and underrepresentation in leadership. Identifying effective measures to mitigate this bias is essential for ensuring fairness in AI recruitment. By evaluating existing strategies, organizations can develop best practices to minimize discrimination and promote equity in hiring.
Data anonymization protects privacy and reduces bias by removing or encrypting personally identifiable information in datasets used for AI training. This helps safeguard privacy while minimizing bias based on sensitive attributes like race, gender, or age. However, balancing privacy with the need for effective AI systems can be challenging. Algorithmic auditing examines AI algorithms for biases and their impact on hiring, ensuring transparency and accountability through regular checks. Fairness-aware machine learning techniques promote equitable hiring by incorporating fairness constraints and optimization during development.
3.4. Exploring laws and regulations to address AI bias in hiring practices
Investigating the impact of laws and regulations on AI and employment discrimination is essential for understanding how the legal system can better protect candidates from biased hiring practices. This research examines past cases, proposed laws, and ethical guidelines, aiming to identify gaps and improve the system.
On September 5, 2024, the United States, the United Kingdom, the European Union, and other nations signed the Artificial Intelligence and Human Rights, Democracy, and the Rule of Law Convention in Vilnius, Lithuania. This treaty addresses AI challenges, promotes responsible innovation, and ensures strong human rights protection. It is the first international agreement to focus on AI’s impact on human rights throughout the AI lifecycle. The convention highlights risks in both public and private sector AI deployment. The legal landscape on AI and employment discrimination is complex, with some countries adopting specific laws and others relying on existing anti-discrimination regulations.
The U.S. legal regulation of AI in recruitment serves as a seminal case. The U.S. Equal Employment Opportunity Commission (EEOC) sued Mackey Education Technology (ITutorGroup) for its online recruitment software's bias against older applicants, violating the Age Discrimination in Employment Act (ADEA). iTutorGroup's software unlawfully filtered out women over 55 and men over 60. The company settled for $365,000 in compensation and implemented anti-discrimination policies and training. This case, which is a legal landmark, shows that employers are still responsible when they use AI to hire people and that discriminatory algorithms could be illegal It portends a wave of litigation involving HR tech and emphasizes the need for companies to ensure their technology adheres to anti-discrimination laws.
The discussion concludes that while AI can revolutionize recruitment, it also brings challenges related to algorithmic bias. Researchers, practitioners, and policymakers must collaborate to ensure ethical AI design and implementation, addressing both technical aspects and broader social and ethical implications.
4. Conclusion
As the paper stands on the precipice of an AI revolution in recruitment, we are faced with a choice: to naively embrace a technology that may entrench inequality or to critically engage with it to ensure it serves as a catalyst for fairness. This paper calls upon all stakeholders to take up the mantle of responsibility and work towards an AI recruitment landscape that is not just efficient but equitable. The question remains: are we ready to hold our creations to the same standards of justice that we hold ourselves?
Artificial intelligence's rapid integration into employment and recruitment has presented both opportunities and challenges. While it promises increased efficiency and a reduction in human biases, the issue of AIHR's algorithmic bias poses a significant threat to fair hiring practices. This paper has endeavored to shed light on the intricacies of algorithmic bias within AI-driven recruitment, its impact on employment discrimination, and the potential avenues for mitigation.
The literature overwhelmingly indicates that AI systems are not inherently neutral but are instead susceptible to the biases present in the data they are trained on. Cases such as Amazon's AI recruitment tool, which exhibited gender bias due to the historical data used in its training, exemplify this. At the technical level, the creation of fair datasets, enhancement of algorithmic transparency, and the implementation of bias-detecting tools are crucial steps towards mitigating bias. However, these measures are not sufficient on their own. There is a pressing need for clear guidelines and regulations that govern the use of AI in recruitment. This includes establishing the responsibilities of recruitment entities and setting standards for accountability and transparency. Additionally, setting up oversight and judging panels can add another level of protection, making sure that AI algorithms are not only fair but also in line with ethical hiring practices. Comprehensive measures are essential to harness AI's potential in recruitment while ensuring fairness and equity. Beyond government control, economic measures like an AI tax could help regulate enterprises by increasing their cost of using AI, indirectly addressing inequity. While not directly solving the problem, this approach reduces the foundation of bias and fosters more equitable development.
In conclusion, the discourse on AI and employment discrimination requires continuous examination and adaptation. As AI technology evolves, so too must our strategies to ensure that it serves as an instrument of empowerment rather than a tool of discrimination. It is incumbent upon all stakeholders, including technologists, HR professionals, policymakers, and society at large, to engage in this conversation and drive the development of AI recruitment tools that are not only efficient but also equitable.
As the author looks to the future, through focused effort and adherence to ethical AI practices, AI in recruitment can become a tool for progress rather than a source of bias.
References
[1]. Jain, L.R., Menon, V. (2023). Algorithmic bias in recruitment: Impacts on racial, gender, and socioeconomic diversity. Journal of Technology and Ethics, 17(3), 45-60.
[2]. Winick, E. (2023) Amazon's AI recruitment tool found to favor male candidates. Archive Page. https://www.example.com/article
[3]. Fabris, A., Baranowska, N., Dennis, M. J., Graus, D., Hacker, P., Saldivar, J. Zuiderveen Borgesius, F. and Biega, A.J. (2024). Fairness and bias in algorithmic hiring: A multidisciplinary survey. ACM Transactions on Intelligent Systems and Technology.
[4]. Ajunwa, I. (2019). The paradox of automation as anti-bias intervention. Cardozo L. Rev., 41, 1671.
[5]. Adams‐Prassl, J., Binns, R., Kelly‐Lyth, A. (2023). Directly discriminatory algorithms. The Modern Law Review, 86(1), 144-175.
[6]. Ali, M., Sapiezynski, P., Bogen, M., Korolova, A., Mislove, A., and Rieke, A. (2019). Discrimination through optimization: How Facebook's Ad delivery can lead to biased outcomes. Proc. ACM Hum. Comput. Interact, 3(CSCW), 1-30.
[7]. Jackson, J. (2021). Algorithmic Bias and Its Impact on Fairness in AI Systems. Journal of Technology and Ethics, 35(4), 123-139.
[8]. Raghavan, M., Barocas, S., Kleinberg, J., and Levy, K.(2020) “Mitigating bias in algorithmic hiring: Evaluating claims and practices”, Proceedings of the Conference on Fairness, Accountability, and Transparency, 481–492.
[9]. Avery, D.R., McKay, P. F., Wilson, D.C. (2023). Gender Stereotypes in Machine Learning Models: Implications for Workplace Recruitment. Journal of Applied Psychology, 108(2), 171-185.
[10]. Beneduce, G. (2020). The Risks of Algorithmic Bias in Hiring: A Case Study of Amazon’s ML-Based Tool. Technology and Society Review, 12(3), 234-248.
[11]. Miasato, L., Silva, M. (2019). Gender Bias in Amazon's Hiring Algorithm: Lessons Learned. Journal of Business Ethics and AI, 8(1), 45-58.
[12]. Langenkamp, A., Berland, A., Sykes, E. (2019). Discriminatory Practices in AI Hiring Systems: A Closer Look at Gender Bias. AI & Society, 34(2), 289-304.
[13]. Faragher, R. (2019). Unintended Consequences of Algorithmic Hiring Tools: The Case of Amazon’s AI System. Journal of Business and Technology Ethics, 7(1), 56-71.
[14]. O'neil, C. (2017). Weapons of math destruction: How big data increases inequality and threatens democracy. Crown.
[15]. Andrews, L., & Bucher, H. (2022). Automating Discrimination: AI Hiring Practices and Gender Inequality. Cardozo L. Rev., 44, 145.
Cite this article
Zhu,Y. (2025). AI and Employment Discrimination: AIHR's Algorithmic Bias. Applied and Computational Engineering,118,89-94.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 3rd International Conference on Software Engineering and Machine Learning
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Jain, L.R., Menon, V. (2023). Algorithmic bias in recruitment: Impacts on racial, gender, and socioeconomic diversity. Journal of Technology and Ethics, 17(3), 45-60.
[2]. Winick, E. (2023) Amazon's AI recruitment tool found to favor male candidates. Archive Page. https://www.example.com/article
[3]. Fabris, A., Baranowska, N., Dennis, M. J., Graus, D., Hacker, P., Saldivar, J. Zuiderveen Borgesius, F. and Biega, A.J. (2024). Fairness and bias in algorithmic hiring: A multidisciplinary survey. ACM Transactions on Intelligent Systems and Technology.
[4]. Ajunwa, I. (2019). The paradox of automation as anti-bias intervention. Cardozo L. Rev., 41, 1671.
[5]. Adams‐Prassl, J., Binns, R., Kelly‐Lyth, A. (2023). Directly discriminatory algorithms. The Modern Law Review, 86(1), 144-175.
[6]. Ali, M., Sapiezynski, P., Bogen, M., Korolova, A., Mislove, A., and Rieke, A. (2019). Discrimination through optimization: How Facebook's Ad delivery can lead to biased outcomes. Proc. ACM Hum. Comput. Interact, 3(CSCW), 1-30.
[7]. Jackson, J. (2021). Algorithmic Bias and Its Impact on Fairness in AI Systems. Journal of Technology and Ethics, 35(4), 123-139.
[8]. Raghavan, M., Barocas, S., Kleinberg, J., and Levy, K.(2020) “Mitigating bias in algorithmic hiring: Evaluating claims and practices”, Proceedings of the Conference on Fairness, Accountability, and Transparency, 481–492.
[9]. Avery, D.R., McKay, P. F., Wilson, D.C. (2023). Gender Stereotypes in Machine Learning Models: Implications for Workplace Recruitment. Journal of Applied Psychology, 108(2), 171-185.
[10]. Beneduce, G. (2020). The Risks of Algorithmic Bias in Hiring: A Case Study of Amazon’s ML-Based Tool. Technology and Society Review, 12(3), 234-248.
[11]. Miasato, L., Silva, M. (2019). Gender Bias in Amazon's Hiring Algorithm: Lessons Learned. Journal of Business Ethics and AI, 8(1), 45-58.
[12]. Langenkamp, A., Berland, A., Sykes, E. (2019). Discriminatory Practices in AI Hiring Systems: A Closer Look at Gender Bias. AI & Society, 34(2), 289-304.
[13]. Faragher, R. (2019). Unintended Consequences of Algorithmic Hiring Tools: The Case of Amazon’s AI System. Journal of Business and Technology Ethics, 7(1), 56-71.
[14]. O'neil, C. (2017). Weapons of math destruction: How big data increases inequality and threatens democracy. Crown.
[15]. Andrews, L., & Bucher, H. (2022). Automating Discrimination: AI Hiring Practices and Gender Inequality. Cardozo L. Rev., 44, 145.