Amodal Instance Segmentation with KINS Dataset

被引:93
作者
Qi, Lu [1 ,2 ]
Jiang, Li [1 ,2 ]
Liu, Shu [2 ]
Shen, Xiaoyong [2 ]
Jia, Jiaya [1 ,2 ]
机构
[1] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[2] Tencent, YouTu Lab, Shenzhen, Peoples R China
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) | 2019年
关键词
D O I
10.1109/CVPR.2019.00313
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Amodal instance segmentation, a new direction of instance segmentation, aims to segment each object instance involving its invisible, occluded parts to imitate human ability. This task requires to reason objects' complex structure. Despite important and futuristic, this task lacks data with large-scale and detailed annotation, due to the difficulty of correctly and consistently labeling invisible parts, which creates the huge barrier to explore the frontier of visual recognition. In this paper, we augment KITTI with more instance pixel-level annotation for 8 categories, which we call KITTI INStance dataset (KINS). We propose the network structure to reason invisible parts via a new multi-task framework with Multi-Level Coding (MLC), which combines information in various recognition levels. Extensive experiments show that our MLC effectively improves both amodal and inmodal segmentation. The KINS dataset and our proposed method are made publicly available.
引用
收藏
页码:3009 / 3018
页数:10
相关论文
共 43 条
  • [1] [Anonymous], 2014, ECCV
  • [2] [Anonymous], 1997, MACHINE LEARNING
  • [3] [Anonymous], 2016, ECCV
  • [4] [Anonymous], 2017, Deformable convolutional networks
  • [5] [Anonymous], 2017, ICCV
  • [6] [Anonymous], 2018, CVPR
  • [7] [Anonymous], 2015, CVPR
  • [8] [Anonymous], 2017, CVPR
  • [9] [Anonymous], 2016, CVPR
  • [10] [Anonymous], 2016, ECCV