Online Learning and Optimization for Revenue Management Problems with Add-on Discounts
David Simchi-Levi (dslevi@mit.edu),
Rui Sun (ruisun@mit.edu) and
Huanan Zhang (huanan.zhang@colorado.edu)
Additional contact information
David Simchi-Levi: Institute for Data, Systems, and Society, Department of Civil & Environmental Engineering, and Operations Research Center, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139
Rui Sun: Institute for Data, Systems, and Society, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139
Huanan Zhang: Leeds School of Business, University of Colorado Boulder, Boulder, Colorado 80309
Management Science, 2022, vol. 68, issue 10, 7402-7421
Abstract:
We study in this paper a revenue-management problem with add-on discounts. The problem is motivated by the practice in the video game industry by which a retailer offers discounts on selected supportive products (e.g., video games) to customers who have also purchased the core products (e.g., video game consoles). We formulate this problem as an optimization problem to determine the prices of different products and the selection of products for add-on discounts. In the base model, we focus on an independent demand structure. To overcome the computational challenge of this optimization problem, we propose an efficient fully polynomial-time approximation scheme (FPTAS) algorithm that solves the problem approximately to any desired accuracy. Moreover, we consider the problem in the setting in which the retailer has no prior knowledge of the demand functions of different products. To solve this joint learning and optimization problem, we propose an upper confidence bound–based learning algorithm that uses the FPTAS optimization algorithm as a subroutine. We show that our learning algorithm can converge to the optimal algorithm that has access to the true demand functions, and the convergence rate is tight up to a certain logarithmic term. We further show that these results for the independent demand model can be extended to multinomial logit choice models. In addition, we conduct numerical experiments with the real-world transaction data we collect from a popular video gaming brand’s online store on Tmall.com. The experiment results illustrate our learning algorithm’s robust performance and fast convergence in various scenarios. We also compare our algorithm with the optimal policy that does not use any add-on discount. The comparison results show the advantages of using the add-on discount strategy in practice.
Keywords: revenue management; add-on discount; online learning; approximation algorithm (search for similar items in EconPapers)
Date: 2022
References: Add references at CitEc
Citations:
Downloads: (external link)
http://dx.doi.org/10.1287/mnsc.2021.4222 (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:inm:ormnsc:v:68:y:2022:i:10:p:7402-7421
Access Statistics for this article
More articles in Management Science from INFORMS Contact information at EDIRC.
Bibliographic data for series maintained by Chris Asher (casher@informs.org).