LOCATE: Localize and Transfer Object Parts for Weakly Supervised Affordance Grounding

被引:13
|
作者
Li, Gen [1 ]
Jampani, Varun [2 ]
Sun, Deqing [2 ]
Sevilla-Lara, Laura [1 ]
机构
[1] Univ Edinburgh, Edinburgh, Midlothian, Scotland
[2] Google Res, Mountain View, CA USA
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
基金
英国工程与自然科学研究理事会;
关键词
D O I
10.1109/CVPR52729.2023.01051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans excel at acquiring knowledge through observation. For example, we can learn to use new tools by watching demonstrations. This skill is fundamental for intelligent systems to interact with the world. A key step to acquire this skill is to identify what part of the object affords each action, which is called affordance grounding. In this paper, we address this problem and propose a framework called LOCATE that can identify matching object parts across images, to transfer knowledge from images where an object is being used (exocentric images used for learning), to images where the object is inactive (egocentric ones used to test). To this end, we first find interaction areas and extract their feature embeddings. Then we learn to aggregate the embeddings into compact prototypes (human, object part, and background), and select the one representing the object part. Finally, we use the selected prototype to guide affordance grounding. We do this in a weakly supervised manner, learning only from image-level affordance and object labels. Extensive experiments demonstrate that our approach outperforms state-of-the-art methods by a large margin on both seen and unseen objects.
引用
收藏
页码:10922 / 10931
页数:10
相关论文
共 50 条
  • [1] Weakly Supervised Multimodal Affordance Grounding for Egocentric Images
    Xu, Lingjing
    Gao, Yang
    Song, Wenfeng
    Hao, Aimin
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 6324 - 6332
  • [2] INTRA: Interaction Relationship-Aware Weakly Supervised Affordance Grounding
    Jang, Ji Ha
    Seo, Hoigi
    Chun, Se Young
    COMPUTER VISION - ECCV 2024, PT LXIV, 2025, 15122 : 18 - 34
  • [3] Weakly Supervised Affordance Detection
    Sawatzky, Johann
    Srikantha, Abhilash
    Gall, Juergen
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5197 - 5206
  • [4] Weakly-Supervised Video Object Grounding via Causal Intervention
    Wang, Wei
    Gao, Junyu
    Xu, Changsheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3933 - 3948
  • [5] Improving Weakly Supervised Scene Graph Parsing through Object Grounding
    Zhang, Yizhou
    Zheng, Zhaoheng
    Nevatia, Ram
    Liu, Yan
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4058 - 4064
  • [6] Weakly-Supervised Video Object Grounding via Stable Context Learning
    Wang, Wei
    Gao, Junyu
    Xu, Changsheng
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 760 - 768
  • [7] Transfer Learning by Ranking for Weakly Supervised Object Annotation
    Shi, Zhiyuan
    Siva, Parthipan
    Xiang, Tao
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2012, 2012,
  • [8] CaT: Weakly Supervised Object Detection with Category Transfer
    Cao, Tianyue
    Du, Lianyu
    Zhang, Xiaoyun
    Chen, Siheng
    Zhang, Ya
    Wang, Yan-Feng
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 3050 - 3059
  • [9] Adaptive Binarization for Weakly Supervised Affordance Segmentation
    Sawatzky, Johann
    Gall, Juergen
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, : 1383 - 1391
  • [10] Weakly-Supervised Video Object Grounding by Exploring Spatio-Temporal Contexts
    Yang, Xun
    Liu, Xueliang
    Jian, Meng
    Gao, Xinjian
    Wang, Meng
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1939 - 1947