Multi-Level Contextual RNNs With Attention Model for Scene Labeling

被引:26
|
作者
Fan, Heng [1 ]
Mei, Xue [2 ]
Prokhorov, Danil [2 ]
Ling, Haibin [1 ]
机构
[1] Temple Univ, Dept Comp & Informat Sci, Philadelphia, PA 19121 USA
[2] Toyota Res Inst, Ann Arbor, MI 48105 USA
关键词
Scene labeling; scene understanding; contextual recurrent neural networks (CRNNs); attention model; intelligent transportation system; FEATURES; SEGMENTATION;
D O I
10.1109/TITS.2017.2775628
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Image context in image is crucial for improving scene labeling. While the existing methods only exploit local context generated from a small surrounding area of an image patch or a pixel, the long-range and global contextual information is often ignored. To handle this issue, we propose a novel approach for scene labeling by multi-level contextual recurrent neural networks (RNNs). We encode three kinds of contextual cues, viz., local context, global context, and image topic context in structural RNNs to model long-range local and global dependencies in an image. In this way, our method is able to "see" the image in terms of both long-range local and holistic views, and make a more reliable inference for image labeling. Besides, we integrate the proposed contextual RNNs into hierarchical convolutional neural networks, and exploit dependence relationships at multiple levels to provide rich spatial and semantic information. Moreover, we adopt an attention model to effectively merge multiple levels and show that it outperforms average- or max-pooling fusion strategies. Extensive experiments demonstrate that the proposed approach achieves improved results on the CamVid, KITTI, SiftFlow, Stanford Background, and Cityscapes data sets.
引用
收藏
页码:3475 / 3485
页数:11
相关论文
共 50 条
  • [31] CNNs with Multi-Level Attention for Domain Generalization
    Ballas, Aristotelis
    Diou, Cristos
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 592 - 596
  • [32] Multi-level attention for referring expression comprehension
    Sun, Yanfeng
    Zhang, Yunru
    Jiang, Huajie
    Hu, Yongli
    Yin, Baocai
    PATTERN RECOGNITION LETTERS, 2023, 172 : 252 - 258
  • [33] Multi-level model of contextual factors and teachers' assessment practices: an integrative review of research
    Fulmer, Gavin
    Lee, Iris
    Tan, Kelvin
    ASSESSMENT IN EDUCATION-PRINCIPLES POLICY & PRACTICE, 2015, 22 (04) : 475 - 494
  • [34] MULTI-LEVEL SEMANTIC LABELING OF SKY/CLOUD IMAGES
    Dev, Soumyabrata
    Lee, Yee Hui
    Winkler, Stefan
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 636 - 640
  • [35] Multi-view Fusion for Multi-level Robotic Scene Understanding
    Lin, Yunzhi
    Tremblay, Jonathan
    Tyree, Stephen
    Vela, Patricio A.
    Birchfield, Stan
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 6817 - 6824
  • [36] Joint Deep Model with Multi-Level Attention and Hybrid-Prediction for Recommendation
    Lin, Zhipeng
    Tang, Yuhua
    Zhang, Yongjun
    ENTROPY, 2019, 21 (02):
  • [37] Road Crack Model Based on Multi-Level Feature Fusion and Attention Mechanism
    Song, Rongrong
    Wang, Caiyong
    Tian, Qichuan
    Zhang, Qi
    Computer Engineering and Applications, 2023, 59 (13): : 281 - 288
  • [38] DTSMLA: A dynamic task scheduling multi-level attention model for stock ranking
    Du, Yuanchuang
    Xie, Liang
    Liao, Sihao
    Chen, Shengshuang
    Wu, Yuchen
    Xu, Haijiao
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 243
  • [39] Lightweight matting method based on multi-level representation cue attention model
    Liu X.-L.
    Zhang L.-C.
    Zhu H.-B.
    Zhang W.-B.
    Kongzhi yu Juece/Control and Decision, 2024, 39 (01): : 87 - 94
  • [40] Research on person re-identification based on multi-level attention model
    Wei, Dan
    Liang, Danyang
    Wu, Longfei
    Wang, Xiaolan
    Jiang, Lei
    Luo, Suyun
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (40) : 87459 - 87477