EconPapers    
Economics at your fingertips  
 

Should I Trust the Artificial Intelligence to Recruit? Recruiters’ Perceptions and Behavior When Faced With Algorithm-Based Recommendation Systems During Resume Screening

Alain Lacroux () and Christelle Martin-Lacroux ()
Additional contact information
Alain Lacroux: LARSH - Laboratoire de Recherche Sociétés & Humanités - UPHF - Université Polytechnique Hauts-de-France - INSA Hauts-De-France - INSA Institut National des Sciences Appliquées Hauts-de-France - INSA - Institut National des Sciences Appliquées
Christelle Martin-Lacroux: CERAG - Centre d'études et de recherches appliquées à la gestion - UGA - Université Grenoble Alpes

Post-Print from HAL

Abstract: Resume screening assisted by decision support systems that incorporate artificial intelligence is currently undergoing a strong development in many organizations, raising technical, managerial, legal, and ethical issues. The purpose of the present paper is to better understand the reactions of recruiters when they are offered algorithm-based recommendations during resume screening. Two polarized attitudes have been identified in the literature on users' reactions to algorithm-based recommendations: algorithm aversion, which reflects a general distrust and preference for human recommendations; and automation bias, which corresponds to an overconfidence in the decisions or recommendations made by algorithmic decision support systems (ADSS). Drawing on results obtained in the field of automated decision support areas, we make the general hypothesis that recruiters trust human experts more than ADSS, because they distrust algorithms for subjective decisions such as recruitment. An experiment on resume screening was conducted on a sample of professionals ( N = 694) involved in the screening of job applications. They were asked to study a job offer, then evaluate two fictitious resumes in a 2 × 2 factorial design with manipulation of the type of recommendation (no recommendation/algorithmic recommendation/human expert recommendation) and of the consistency of the recommendations (consistent vs. inconsistent recommendation). Our results support the general hypothesis of preference for human recommendations: recruiters exhibit a higher level of trust toward human expert recommendations compared with algorithmic recommendations. However, we also found that recommendation's consistence has a differential and unexpected impact on decisions: in the presence of an inconsistent algorithmic recommendation, recruiters favored the unsuitable over the suitable resume. Our results also show that specific personality traits (extraversion, neuroticism, and self-confidence) are associated with a differential use of algorithmic recommendations. Implications for research and HR policies are finally discussed.

Date: 2022-07-06
Note: View the original document on HAL open archive server: https://paris1.hal.science/hal-04011972v1
References: Add references at CitEc
Citations:

Published in Frontiers in Psychology, 2022, 13, ⟨10.3389/fpsyg.2022.895997⟩

Downloads: (external link)
https://paris1.hal.science/hal-04011972v1/document (application/pdf)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:hal:journl:hal-04011972

DOI: 10.3389/fpsyg.2022.895997

Access Statistics for this paper

More papers in Post-Print from HAL
Bibliographic data for series maintained by CCSD ().

 
Page updated 2025-03-19
Handle: RePEc:hal:journl:hal-04011972