VFM: Visual Feedback Model for Robust Object Recognition

被引:0
|
作者
Chong Wang
Kai-Qi Huang
机构
[1] Institute of Automation,National Laboratory of Pattern Recognition
[2] Chinese Academy of Sciences,undefined
来源
Journal of Computer Science and Technology | 2015年 / 30卷
关键词
object recognition; object classification; object detection; visual feedback;
D O I
暂无
中图分类号
学科分类号
摘要
Object recognition, which consists of classification and detection, has two important attributes for robustness: 1) closeness: detection windows should be as close to object locations as possible, and 2) adaptiveness: object matching should be adaptive to object variations within an object class. It is difficult to satisfy both attributes using traditional methods which consider classification and detection separately; thus recent studies propose to combine them based on confidence contextualization and foreground modeling. However, these combinations neglect feature saliency and object structure, and biological evidence suggests that the feature saliency and object structure can be important in guiding the recognition from low level to high level. In fact, object recognition originates in the mechanism of “what” and “where” pathways in human visual systems. More importantly, these pathways have feedback to each other and exchange useful information, which may improve closeness and adaptiveness. Inspired by the visual feedback, we propose a robust object recognition framework by designing a computational visual feedback model (VFM) between classification and detection. In the “what” feedback, the feature saliency from classification is exploited to rectify detection windows for better closeness; while in the “where” feedback, object parts from detection are used to match object structure for better adaptiveness. Experimental results show that the “what” and “where” feedback is effective to improve closeness and adaptiveness for object recognition, and encouraging improvements are obtained on the challenging PASCAL VOC 2007 dataset.
引用
收藏
页码:325 / 339
页数:14
相关论文
共 50 条
  • [31] Robust model-based object recognition using a dual-hierarchy graph
    Weiss, Isaac
    AUTOMATIC TARGET RECOGNITION XXII, 2012, 8391
  • [32] Compositional Convolutional Neural Networks: A Robust and Interpretable Model for Object Recognition Under Occlusion
    Adam Kortylewski
    Qing Liu
    Angtian Wang
    Yihong Sun
    Alan Yuille
    International Journal of Computer Vision, 2021, 129 : 736 - 760
  • [33] Compositional Convolutional Neural Networks: A Robust and Interpretable Model for Object Recognition Under Occlusion
    Kortylewski, Adam
    Liu, Qing
    Wang, Angtian
    Sun, Yihong
    Yuille, Alan
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (03) : 736 - 760
  • [34] Omnidirectional Image Stabilization for Visual Object Recognition
    Akihiko Torii
    Michal Havlena
    Tomáš Pajdla
    International Journal of Computer Vision, 2011, 91 : 157 - 174
  • [35] Distributed Object Recognition in Visual Sensor Networks
    Paris, Stefano
    Redondi, Alessandro
    Cesana, Matteo
    Tagliasacchi, Marco
    2015 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2015, : 6701 - 6706
  • [36] Exploiting Object Similarity for Robotic Visual Recognition
    Cai, Hong
    Mostofi, Yasamin
    IEEE TRANSACTIONS ON ROBOTICS, 2021, 37 (01) : 16 - 33
  • [37] Exploiting Core Knowledge for Visual Object Recognition
    Schurgin, Mark W.
    Flombaum, Jonathan I.
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-GENERAL, 2017, 146 (03) : 362 - 375
  • [38] Unsupervised Learning of Visual Object Recognition Models
    Navarrete, Dulce J.
    Morales, Eduardo F.
    Enrique Sucar, Luis
    ADVANCES IN ARTIFICIAL INTELLIGENCE - IBERAMIA 2012, 2012, 7637 : 511 - 520
  • [39] IDENTIFYING AND LEARNING VISUAL ATTRIBUTES FOR OBJECT RECOGNITION
    Wan, Kong-Wah
    Roy, Sujoy
    2010 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 2010, : 3893 - 3896
  • [40] A restricted 'spotlight' of attention in visual object recognition
    Stark, ME
    Grafman, J
    Fertig, E
    NEUROPSYCHOLOGIA, 1997, 35 (09) : 1233 - 1249