GENERATING COHERENT NATURAL LANGUAGE ANNOTATIONS FOR VIDEO STREAMS

被引:0
|
作者
Khan, Muhammad Usman Ghani [1 ]
Zhang, Lei [2 ]
Gotoh, Yoshihiko [1 ]
机构
[1] Univ Sheffield, Sheffield, S Yorkshire, England
[2] Harbin Engn Univ, Harbin, Heilongjiang, Peoples R China
来源
2012 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2012) | 2012年
关键词
Video processing; Video annotation; Natural language description; video feature units;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
This contribution addresses generation of natural language annotations for human actions, behaviour and their interactions with other objects observed in video streams. The work starts with implementation of conventional image processing techniques to extract high level features for individual frames. Natural language description of the frame contents is produced based on high level features. Although feature extraction processes are erroneous at various levels, we explore approaches to put them together to produce a coherent description. For extending the approach to description of video streams, units of features are introduced to present coherent, smooth and well phrased descriptions by incorporating spatial and temporal information. Evaluation is made by calculating ROUGE scores between human annotated and machine generated descriptions.
引用
收藏
页码:2893 / 2896
页数:4
相关论文
共 47 条
  • [1] A framework for creating natural language descriptions of video streams
    Khan, Muhammad Usman Ghani
    Al Harbi, Nouf
    Gotoh, Yoshihiko
    INFORMATION SCIENCES, 2015, 303 : 61 - 82
  • [2] Generating natural language tags for video information management
    Muhammad Usman Ghani Khan
    Yoshihiko Gotoh
    Machine Vision and Applications, 2017, 28 : 243 - 265
  • [3] Generating natural language tags for video information management
    Khan, Muhammad Usman Ghani
    Gotoh, Yoshihiko
    MACHINE VISION AND APPLICATIONS, 2017, 28 (3-4) : 243 - 265
  • [4] Literature review on video annotations in teacher education
    Cebrian-Robles, Violeta
    Perez-Torregrosa, Ana-Belen
    de la Serna, Manuel Cebrian
    PIXEL-BIT- REVISTA DE MEDIOS Y EDUCACION, 2023, (66): : 31 - 57
  • [5] Learning Physics Through Online Video Annotations
    Marcal, J.
    Borges, M. M.
    Viana, P.
    Carvalho, P.
    EDUCATION IN THE KNOWLEDGE SOCIETY, 2020, 21
  • [6] Cluster Analysis of Learners based on Video Annotations
    Furukawa, Masako
    Yaginuma, Yoshitomo
    Yamada, Tsuneo
    2018 7TH INTERNATIONAL CONGRESS ON ADVANCED APPLIED INFORMATICS (IIAI-AAI 2018), 2018, : 944 - 945
  • [7] AffecTube - Chrome extension for YouTube video affective annotations
    Kulas, Daniel
    Wrobel, Michal R.
    SOFTWAREX, 2023, 23
  • [8] Linking User Generated Video Annotations to the Web of Data
    Hildebrand, Michiel
    van Ossenbruggen, Jacco
    ADVANCES IN MULTIMEDIA MODELING, 2012, 7131 : 693 - 704
  • [9] Annotations for streaming video on the Web: system design and usage studies
    Bargeron, D
    Gupta, A
    Grudin, J
    Sanocki, E
    PROCEEDINGS OF THE EIGHTH INTERNATIONAL WORLD WIDE WEB CONFERENCE, 1999, : 61 - 75
  • [10] Annotations for streaming video on the Web: system design and usage studies
    Bargeron, D
    Gupta, A
    Grudin, J
    Sanocki, E
    COMPUTER NETWORKS, 1999, 31 (11-16) : 1139 - 1153