Modelling perceptions on the evaluation of video summarization

被引:2
作者
Abdalla, Kalyf [1 ,2 ]
Menezes, Igor [3 ]
Oliveira, Luciano [1 ]
机构
[1] Univ Fed Bahia, Intelligent Vis Res Lab, Salvador, BA, Brazil
[2] Fed Inst Bahia, Salvador, BA, Brazil
[3] Univ Hull, Kingston Upon Hull, N Humberside, England
关键词
Video summarization; Subjective evaluation; Evaluation metric; SHOT-BOUNDARY DETECTION; RETRIEVAL;
D O I
10.1016/j.eswa.2019.04.065
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hours of video are uploaded to streaming platforms every minute, with recommender systems suggesting popular and relevant videos that can help users save time in the searching process. Recommender systems regularly require video summarization as an expert system to automatically identify suitable video entities and events. Since there is no well-established methodology to evaluate the relevance of summarized videos, some studies have made use of user annotations to gather evidence about the effectiveness of summarization methods. Aimed at modelling the user's perceptions, which ultimately form the basis for testing video summarization systems, this paper seeks to propose: (i) A guideline to collect unrestricted user annotations, (ii) a novel metric called compression level of user annotation (CLUSA) to gauge the performance of video summarization methods, and (iii) a study on the quality of annotated video summaries collected from different assessment scales. These contributions lead to benchmarking video summarization methods with no constraints, even if user annotations are collected from different assessment scales for each method. Our experiments showed that CLUSA is less susceptible to unbalanced compression data sets in comparison to other metrics, hence achieving higher reliability estimates. CLUSA also allows to compare results from different video summarizing approaches. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:254 / 265
页数:12
相关论文
共 40 条
[1]   Design and evaluation of a music video summarization system [J].
Agnihotri, L ;
Dimitrova, N ;
Kender, JR .
2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3, 2004, :1943-1946
[2]  
[Anonymous], 2012, P SIGCHI C HUM FACT, DOI [DOI 10.1145/2207676.2207767, 10.1145/2207676.22077672, DOI 10.1145/2207676.22077672]
[3]  
[Anonymous], 2012, UCF101 DATASET 101 H
[4]  
[Anonymous], 2017, TREC VIDEO RETRIEVAL
[5]  
[Anonymous], TECHNICAL REPORT
[6]  
Arman E., 1994, Proceedings ACM Multimedia '94, P97, DOI 10.1145/192593.192630
[7]   Instance search retrospective with focus on TRECVID [J].
Awad G. ;
Kraaij W. ;
Over P. ;
Satoh S. .
International Journal of Multimedia Information Retrieval, 2017, 6 (1) :1-29
[8]  
Chang P., 2002, P INT C IM PROC, DOI 10.1109/1C1P.2002.1038097
[9]  
Chu WS, 2015, PROC CVPR IEEE, P3584, DOI 10.1109/CVPR.2015.7298981
[10]   Video Summarization via Segments Summary Graphs [J].
Demir, Mahmut ;
Bozma, H. Isil .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOP (ICCVW), 2015, :1071-1077