A Theory-Based Explainable Deep Learning Architecture for Music Emotion

被引:2
作者
Fong, Hortense [1 ]
Kumar, Vineet [2 ]
Sudhir, K. [2 ]
机构
[1] Columbia Business Sch, Mkt, New York, NY 10027 USA
[2] Yale Sch Management, New Haven, CT 06511 USA
关键词
audio data; deep learning; explainable and interpretable AI; emotion; digital advertising; music theory; RESPONSES; DISCRETE; CONTEXT; MODEL; MOOD; FELT;
D O I
10.1287/mksc.2022.0323
中图分类号
F [经济];
学科分类号
02 ;
摘要
This paper develops a theory-based, explainable deep learning convolutional neural network (CNN) classifier to predict the time-varying emotional response to music. We design novel CNN filters that leverage the frequency harmonics structure from acoustic physics known to impact the perception of musical features. Our theory-based model is more parsimonious, but it provides comparable predictive performance with atheoretical deep learning models while performing better than models using handcrafted features. Our model can be complemented with handcrafted features, but the performance improvement is marginal. Importantly, the harmonics-based structure placed on the CNN filters provides better explainability for how the model predicts emotional response (valence and arousal) because emotion is closely related to consonance-a perceptual feature defined by the alignment of harmonics. Finally, we illustrate the utility of our model with an application involving digital advertising. Motivated by YouTube's midroll ads, we conduct a laboratory experiment in which we exogenously insert ads at different times within videos. We find that ads placed in emotionally similar contexts increase ad engagement (lower skip rates and higher brand recall rates). Ad insertion based on emotional similarity metrics predicted by our theory-based, explainable model produces comparable or better engagement relative to atheoretical models.
引用
收藏
页码:196 / 219
页数:25
相关论文
共 73 条
[1]   Developing a benchmark for emotional analysis of music [J].
Aljanaki, Anna ;
Yang, Yi-Hsuan ;
Soleymani, Mohammad .
PLOS ONE, 2017, 12 (03)
[2]   A content analysis of music placement in prime-time television advertising [J].
Allan, David .
JOURNAL OF ADVERTISING RESEARCH, 2008, 48 (03) :404-417
[3]   Behavioral consequences of affect: Combining evaluative and regulatory mechanisms [J].
Andrade, EB .
JOURNAL OF CONSUMER RESEARCH, 2005, 32 (03) :355-362
[4]  
[Anonymous], 2006, The Cambridge History of Western Music Theory
[5]   Understanding Interactive Online Advertising: Congruence and Product Involvement in Highly and Lowly Arousing, Skippable Video Ads [J].
Belanche, Daniel ;
Flavian, Carlos ;
Perez-Rueda, Alfredo .
JOURNAL OF INTERACTIVE MARKETING, 2017, 37 :75-88
[6]   Musical Instrument Identification Using Deep Learning Approach [J].
Blaszke, Maciej ;
Kostek, Bozena .
SENSORS, 2022, 22 (08)
[7]   Dynamics of Musical Success: A Machine Learning Approach for Multimedia Data Fusion [J].
Boughanmi, Khaled ;
Ansari, Asim .
JOURNAL OF MARKETING RESEARCH, 2021, 58 (06) :1034-1057
[8]   A Close Look at Deep Learning with Small Data [J].
Brigato, Lorenzo ;
Iocchi, Luca .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :2490-2497
[9]   MUSIC, MOOD, AND MARKETING [J].
BRUNER, GC .
JOURNAL OF MARKETING, 1990, 54 (04) :94-104
[10]  
Bullerjahn C., 1994, Psychomusicology, V13, P99, DOI DOI 10.1037/H0094100