Affective video content representation and modeling

被引:353
|
作者
Hanjalic, A [1 ]
Xu, LQ
机构
[1] Delft Univ Technol, Dept Mediamat, NL-2628 CD Delft, Netherlands
[2] Martlesham Hlth, BT Res Venturing, Broadband Appl Res Ctr, Ipswich IP5 3RE, Suffolk, England
关键词
affective video content analysis; video abstraction; video content modeling; video content representation; video highlights extraction;
D O I
10.1109/TMM.2004.840618
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper looks into a new direction in video content analysis - the representation and modeling of affective video content. The affective content of a given video clip can be defined as the intensity and type of feeling or emotion (both are referred to as affect) that are expected to arise in the user while watching that clip. The availability of methodologies for automatically extracting this type of video content will extend the current scope of possibilities for video indexing and retrieval. For instance, we will be able to search for the funniest or the most thrilling parts of a movie, or the most exciting events of a sport program. Furthermore, as the user may want to select a movie not only based on its genre, cast, director and story content, but also on its prevailing mood, the affective content analysis is also likely to contribute to enhancing the quality of personalizing the video delivery to the user. We propose in this paper a computational framework for affective video content representation and modeling. This framework is based on the dimensional approach to affect that is known from the field of psychophysiology. According to this approach, the affective video content can be represented as a set of points in the two-dimensional (2-D) emotion space that is characterized by the dimensions of arousal (intensity of affect) and valence (type of affect). We map the affective video content onto the 2-D emotion space by using the models that link the arousal and valence dimensions to low-level features extracted from video data. This results in the arousal and valence time curves that, either considered separately or combined into the so-called affect curve, are introduced as reliable representations of expected transitions from one feeling to another along a video, as perceived by a viewer.
引用
收藏
页码:143 / 154
页数:12
相关论文
共 50 条
  • [2] Video affective content representation and recognition using video affective tree and Hidden Markov Models
    Sun, Kai
    Yu, Junqing
    AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, PROCEEDINGS, 2007, 4738 : 594 - 605
  • [3] AN IMPROVED VALENCE-AROUSAL EMOTION SPACE FOR VIDEO AFFECTIVE CONTENT REPRESENTATION AND RECOGNITION
    Sun, Kai
    Yu, Junqing
    Huang, Yue
    Hu, Xiaoqiang
    ICME: 2009 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-3, 2009, : 566 - 569
  • [4] Effective content representation for video
    Ferman, AM
    Tekalp, AM
    Mehrotra, R
    1998 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING - PROCEEDINGS, VOL 3, 1998, : 521 - 525
  • [5] Towards affective level video applications: A novel FPGA-based video arousal content modeling system
    Arifin, Sutjipto
    Chang, Peter Y. K.
    2006 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS, PROCEEDINGS, 2006, : 713 - 716
  • [6] Representation Learning through Multimodal Attention and Time-Sync Comments for Affective Video Content Analysis
    Pan, Jicai
    Wang, Shangfei
    Fang, Lin
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022,
  • [7] Affective Video Content Analysis: A Multidisciplinary Insight
    Baveye, Yoann
    Chamaret, Christel
    Dellandrea, Emmanuel
    Chen, Liming
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2018, 9 (04) : 396 - 409
  • [8] Learning Affective Features Based on VIP for Video Affective Content Analysis
    Zhu, Yingying
    Tong, Min
    Huang, Tinglin
    Wen, Zhenkun
    Tian, Qi
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT III, 2018, 11166 : 697 - 707
  • [9] Video content representation on tiny devices
    Wang, J
    Reinders, MJT
    Lagendijk, RL
    Lindenberg, J
    Kankanhalli, MS
    2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3, 2004, : 1711 - 1714
  • [10] Audio-visual and EEG-based Attention Modeling for Extraction of Affective Video Content
    Mehmood, Irfan
    Sajjad, Muhammad
    Baik, Sung Wook
    Rho, Seungmin
    2015 INTERNATIONAL CONFERENCE ON PLATFORM TECHNOLOGY AND SERVICE (PLATCON), 2015, : 17 - 18