Learning Personalized Product Recommendations with Customer Disengagement
Hamsa Bastani (),
Pavithra Harsha (),
Georgia Perakis () and
Divya Singhvi ()
Additional contact information
Hamsa Bastani: Operations Information and Decisions, Wharton School, University of Pennsylvania, Philadelphia, Pennsylvania 19104
Pavithra Harsha: IBM Thomas J. Watson Research Center, Yorktown Heights, New York 10598
Georgia Perakis: Sloan School of Management, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139
Divya Singhvi: Leonard N. Stern School of Business, New York University, New York, New York 10012
Manufacturing & Service Operations Management, 2022, vol. 24, issue 4, 2010-2028
Abstract:
Problem definition : We study personalized product recommendations on platforms when customers have unknown preferences. Importantly, customers may disengage when offered poor recommendations. Academic/practical relevance : Online platforms often personalize product recommendations using bandit algorithms, which balance an exploration-exploitation trade-off. However, customer disengagement—a salient feature of platforms in practice—introduces a novel challenge because exploration may cause customers to abandon the platform. We propose a novel algorithm that constrains exploration to improve performance. Methodology : We present evidence of customer disengagement using data from a major airline’s ad campaign; this motivates our model of disengagement, where a customer may abandon the platform when offered irrelevant recommendations. We formulate the customer preference learning problem as a generalized linear bandit, with the notable difference that the customer’s horizon length is a function of past recommendations. Results : We prove that no algorithm can keep all customers engaged. Unfortunately, classical bandit algorithms provably overexplore, causing every customer to eventually disengage. Motivated by the structural properties of the optimal policy in a scalar instance of our problem, we propose modifying bandit learning strategies by constraining the action space up front using an integer program. We prove that this simple modification allows our algorithm to perform well by keeping a significant fraction of customers engaged. Managerial implications : Platforms should be careful to avoid overexploration when learning customer preferences if customers have a high propensity for disengagement. Numerical experiments on movie recommendations data demonstrate that our algorithm can significantly improve customer engagement.
Keywords: bandits; recommendation systems; collaborative filtering; disengagement; cold start (search for similar items in EconPapers)
Date: 2022
References: Add references at CitEc
Citations:
Downloads: (external link)
http://dx.doi.org/10.1287/msom.2021.1047 (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:inm:ormsom:v:24:y:2022:i:4:p:2010-2028
Access Statistics for this article
More articles in Manufacturing & Service Operations Management from INFORMS Contact information at EDIRC.
Bibliographic data for series maintained by Chris Asher ().