When consumers need more interpretability of artificial intelligence (AI) recommendations? The effect of decision-making domains
Changdong Chen and
Yuchen Zheng
Behaviour and Information Technology, 2024, vol. 43, issue 14, 3481-3489
Abstract:
Due to the “black-box’ nature of artificial intelligence (AI) recommendations, interpretability is critical to the consumer experience of human-AI interaction. Unfortunately, improving the interpretability of AI recommendations is technically challenging and costly. Therefore, there is an urgent need for the industry to identify when the interpretability of AI recommendations is more likely to be needed. This study defines the construct of Need for Interpretability (NFI) of AI recommendations and empirically tests consumers’ need for interpretability of AI recommendations in different decision-making domains. Across two experimental studies, we demonstrate that consumers do indeed have a need for interpretability toward AI recommendations, and that the need for interpretability is higher in utilitarian domains than in hedonic domains. This study would help companies to identify the varying need for interpretability of AI recommendations in different application scenarios.
Date: 2024
References: Add references at CitEc
Citations:
Downloads: (external link)
http://hdl.handle.net/10.1080/0144929X.2023.2279658 (text/html)
Access to full text is restricted to subscribers.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:taf:tbitxx:v:43:y:2024:i:14:p:3481-3489
Ordering information: This journal article can be ordered from
http://www.tandfonline.com/pricing/journal/tbit20
DOI: 10.1080/0144929X.2023.2279658
Access Statistics for this article
Behaviour and Information Technology is currently edited by Dr Panos P Markopoulos
More articles in Behaviour and Information Technology from Taylor & Francis Journals
Bibliographic data for series maintained by Chris Longhurst ().