A Multiarmed Bandit Approach for House Ads Recommendations

被引:7
作者
Aramayo, Nicolas [1 ]
Schiappacasse, Mario [1 ]
Goic, Marcel [2 ]
机构
[1] I2B Technol, Santiago 7500010, Chile
[2] Univ Chile, Dept Ind Engn, Santiago 8370456, Chile
关键词
multiarmed bandits; house ads; personalization; deep learning; DYNAMIC ASSORTMENT; NEURAL-NETWORKS; CHOICE; ACQUISITION; DROPOUT; CONTEXT; DESIGN;
D O I
10.1287/mksc.2022.1378
中图分类号
F [经济];
学科分类号
02 ;
摘要
Nowadays, websites use a variety of recommendation systems to decide the content to display to their visitors. In this work, we use a multiarmed bandit approach to dynamically select the combination of house ads to exhibit to a heterogeneous set of customers visiting the website of a large retailer. House ads correspond to promotional information displayed on the website to highlight some specific products and are an important marketing tool for online retailers. As the number of clicks they receive not only depends on their own attractiveness but also on how attractive are other products displayed around them, we decide about complete collections of ads that capture those interactions. Moreover, as ads can wear out, in our recommendations we allow for nonstationary rewards. Furthermore, considering the sparsity of customer-level information, we embed a deep neural network to provide personalized recommendations within a bandit scheme. We tested our methods in controlled experiments where we compared them against decisions made by an experienced team of managers and the recommendations of a variety of other bandit policies. Our results show a more active exploration of the decision space and a significant increment in click-through and add-to-cart rates.
引用
收藏
页码:271 / 292
页数:22
相关论文
共 83 条
[1]   Effective Evaluation Using Logged Bandit Feedback from Multiple Loggers [J].
Agarwal, Aman ;
Basu, Soumya ;
Schnabel, Tobias ;
Joachims, Thorsten .
KDD'17: PROCEEDINGS OF THE 23RD ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2017, :687-696
[2]  
Agrawal P, 2020, PREPRINT
[3]  
Agrawal R., 2013, International WorldWideWeb Conference Committee (IW3C2), P13, DOI DOI 10.1145/2488388.2488391
[4]  
Agrawal S., 2012, JMLR WORKSHOP C P, P39
[5]  
Agrawal S., 2013, PMLR, P127, DOI DOI 10.5555/3042817.3043073
[6]   MNL-Bandit: A Dynamic Learning Approach to Assortment Selection [J].
Agrawal, Shipra ;
Avadhanula, Vashist ;
Goyal, Vineet ;
Zeevi, Assaf .
OPERATIONS RESEARCH, 2019, 67 (05) :1453-1485
[7]  
[Anonymous], 2013, PMLR
[8]  
Awerbuch Baruch, 2004, P 36 ANN ACM S THEOR, P45
[9]   Information acquisition and efficient mechanism design [J].
Bergemann, D ;
Välimäki, J .
ECONOMETRICA, 2002, 70 (03) :1007-1033
[10]  
Bergemann D, 2005, RAND J ECON, V36, P719