Object affordance detection with boundary-preserving network for robotic manipulation tasks

被引:0
作者
Congcong Yin
Qiuju Zhang
机构
[1] Jiangnan University,School of Mechanical Engineering
[2] Jiangnan University,Jiangsu Province Key Laboratory of Advanced Food Manufacturing Equipment and Technology
来源
Neural Computing and Applications | 2022年 / 34卷
关键词
Object affordance detection; Instance segmentation; Attention mechanism; Robotic manipulation task;
D O I
暂无
中图分类号
学科分类号
摘要
Object affordance detection aims to identify, locate and segment the functional regions of objects, so that robots can understand and manipulate objects like humans. The affordance detection task has two main challenges: (1) Due to the need to provide accurate positioning information for the robot to manipulate objects, the affordance segmentation results are required to have high boundary quality. (2) Different kinds of objects have significant differences in appearances, but may have the same affordance. Correspondingly, parts with the same appearance may have different affordances. The existing methods regard affordance detection as an image segmentation problem, without focusing on the boundary quality of detection results. In addition, most of the existing methods do not consider the potential relationship between object categories and object affordances. Aiming at the above problems, we propose a boundary-preserving network (BPN) to provide affordance masks with better boundary quality for robots to manipulate objects. Our framework contains three new components: the IoU (Intersection-over-Union) branch, the affordance boundary branch and the relationship attention module. The IoU branch is used to predict the IoU score of each object bounding box. The affordance boundary branch is used to guide the network to learn the boundary features of objects. The relationship attention module is used to enhance the feature representation capability of the network by exploring the potential relationship between object categories and object affordances. Experiments show that our method is helpful to improve the boundary quality of the predicted affordance masks. On the IIT-AFF dataset, the performance of the proposed BPN is 2.32% (F-score) and 2.89% (F-score) higher than that of the strong baseline in terms of affordance masks and the boundaries of affordance masks, respectively. Furthermore, the real-world robot manipulation experiments show that the proposed BPN can provide accurate affordance information for robots to manipulate objects.
引用
收藏
页码:17963 / 17980
页数:17
相关论文
共 69 条
  • [1] Hassanin M(2021)Visual affordance and function understanding: a survey Acm Comput Surv 4 1140-1147
  • [2] Khan S(2019)Learning affordance segmentation for real-world robotic manipulation via synthetic images IEEE Robot Autom Lett 32 14321-14333
  • [3] Tahtali M(2020)Object affordance detection with relationship-aware network Neural Comput Appl 440 36-44
  • [4] Chu F(2021)Visual affordance detection using an efficient attention convolutional neural network Neurocomputing 42 386-397
  • [5] Xu R(2020)Mask R-CNN IEEE Trans Pattern Anal Mach Intell 39 1137-1149
  • [6] Vela PA(2017)Faster R-CNN: towards real-time object detection with region proposal networks IEEE Trans Pattern Anal Mach Intell 100 145-163
  • [7] Zhao X(2021)A study on segmentation and refinement of key human body parts by integrating manual measurements Ergonomics 24 15-26
  • [8] Cao Y(2020)Geometric affordance perception: leveraging deep 3D saliency with the interaction tensor Front Neurorobot 13 798-809
  • [9] Kang Y(2020)Grasp pose detection with affordance-based task constraint learning in single-view point clouds J Intell Rob Syst 3 3465-3472
  • [10] Gu QP(2008)Learning object affordances: from sensory-motor coordination to imitation IEEE Trans Rob 43 1155-1172