Causal Inference with Knowledge Distilling and Curriculum Learning for Unbiased VQA

被引:27
|
作者
Pan, Yonghua [1 ]
Li, Zechao [1 ]
Zhang, Liyan [2 ]
Tang, Jinhui [1 ]
机构
[1] Nanjing Univ Sci & Technol, 200 Xiaolingwei St, Nanjing 210094, Jiangsu, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut, 29 Yudao St, Nanjing 210016, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Visual question answering; neural networks; knowledge distillation; causal inference;
D O I
10.1145/3487042
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, many Visual Question Answering (VQA) models rely on the correlations between questions and answers yet neglect those between the visual information and the textual information. They would perform badly if the handled data distribute differently from the training data (i.e., out-of-distribution (OOD) data). Towards this end, we propose a two-stage unbiased VQA approach that addresses the unbiased issue from a causal perspective. In the causal inference stage, we mark the spurious correlation on the causal graph, explore the counterfactual causality, and devise a causal target based on the inherent correlations between the conventional and counterfactual VQA models. In the distillation stage, we introduce the causal target into the training process and leverages distilling as well as curriculum learning to capture the unbiased model. Since Causal Inference with Knowledge Distilling and Curriculum Learning (CKCL) reinforces the contribution of the visual information and eliminates the impact of the spurious correlation by distilling the knowledge in causal inference to the VQA model, it contributes to the good performance on both the standard data and out-of-distribution data. The extensive experimental results on VQA-CP v2 dataset demonstrate the superior performance of the proposed method compared to the state-of-the-art (SotA) methods.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Distilling Knowledge Based on Curriculum Learning for Temporal Knowledge Graph Embeddings
    Zhang, Bin
    Li, Jiayin
    Dai, Yuanfei
    PROCEEDINGS OF THE 33RD ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2024, 2024, : 4248 - 4252
  • [2] Distilling Knowledge in Federated Learning
    Le, Huy Q.
    Shin, Jong Hoon
    Nguyen, Minh N. H.
    Hong, Choong Seon
    2021 22ND ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS), 2021, : 196 - 201
  • [3] Unbiased Learning for the Causal Effect of Recommendation
    Sato, Masahiro
    Takemori, Sho
    Singh, Janmajay
    Ohkuma, Tomoko
    RECSYS 2020: 14TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, 2020, : 378 - 387
  • [4] Distilling Knowledge on Text Graph for Social Media Attribute Inference
    Li, Quan
    Li, Xiaoting
    Chen, Lingwei
    Wu, Dinghao
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2024 - 2028
  • [5] Single-Head Lifelong Learning Based on Distilling Knowledge
    Wang, Yen-Hsiang
    Lin, Chih-Yang
    Thaipisutikul, Tipajin
    Shih, Timothy K.
    IEEE ACCESS, 2022, 10 : 35469 - 35478
  • [6] Causality, causal discovery, causal inference and counterfactuals in Civil Engineering: Causal machine learning and case studies for knowledge discovery
    Naser, M. Z.
    Tapeh, Arash Teymori Gharah
    COMPUTERS AND CONCRETE, 2023, 31 (04) : 277 - 292
  • [7] Causal discovery and inference for evaluating fire resistance of structural members through causal learning and domain knowledge
    Naser, M. Z.
    Ciftcioglu, Aybike Ozyuksel
    STRUCTURAL CONCRETE, 2023, 24 (03) : 3314 - 3328
  • [8] Evaluating Uses of Deep Learning Methods for Causal Inference
    Whata, Albert
    Chimedza, Charles
    IEEE ACCESS, 2022, 10 : 2813 - 2827
  • [9] Self-supervised knowledge distillation in counterfactual learning for VQA
    Bi, Yandong
    Jiang, Huajie
    Zhang, Hanfu
    Hu, Yongli
    Yin, Baocai
    PATTERN RECOGNITION LETTERS, 2024, 177 : 33 - 39
  • [10] Stable Knowledge Tracing Using Causal Inference
    Zhu, Jia
    Ma, Xiaodong
    Huang, Changqin
    IEEE TRANSACTIONS ON LEARNING TECHNOLOGIES, 2024, 17 : 124 - 134