Cross-Modal Meta-Knowledge Transfer: A Meta-Learning Framework Adaptable for Multimodal Tasks

被引:0
|
作者
Chen, Yuhe [1 ]
Jin, Jingxuan [1 ]
Li, De [1 ]
Wang, Peng [1 ]
机构
[1] Yanbian Univ, Coll Engn, Yanji 133000, Peoples R China
来源
PROCEEDINGS OF 2024 INTERNATIONAL CONFERENCE ON COMPUTER AND MULTIMEDIA TECHNOLOGY, ICCMT 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Few-shot learning; Multimodal learning; Meta-learning;
D O I
10.1145/3675249.3675347
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Due to the significant disparities among different modal data, multimodal few-shot learning has always been a challenging issue in the field of artificial intelligence. Compared to traditional machine learning, meta-learning, as a more data-efficient training framework, its application in multimodal few-shot tasks has not yet been thoroughly investigated. For this reason, this paper proposes a novel two-stage multimodal meta-learning framework. Specifically, we first define the construction method of multimodal meta-tasks, decomposing the model's training into a series of multimodal meta-task collections. This framework actively learns the complementary information of different modalities in a phased manner. Secondly, by acquiring additional textual information as training samples through a language generation model and combining it with images, the multimodal semantic features extracted by the model are enriched. Lastly, we evaluated our method using few-shot classification tasks. Experimental results indicate that our proposed training framework surpasses other methods in recent years across multiple datasets.
引用
收藏
页码:558 / 563
页数:6
相关论文
共 50 条
  • [1] Deep Multimodal Transfer Learning for Cross-Modal Retrieval
    Zhen, Liangli
    Hu, Peng
    Peng, Xi
    Goh, Rick Siow Mong
    Zhou, Joey Tianyi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (02) : 798 - 810
  • [2] TOOLS AND TASKS FOR LEARNING AND META-LEARNING
    Jaworski, Barbara
    JOURNAL OF MATHEMATICS TEACHER EDUCATION, 2005, 8 (05) : 359 - 361
  • [3] Tools and Tasks for Learning and Meta-learning
    Barbara Jaworski
    Journal of Mathematics Teacher Education, 2005, 8 (5) : 359 - 361
  • [4] Cross-Modal Dynamic Transfer Learning for Multimodal Emotion Recognition
    Hong, Soyeon
    Kang, Hyeoungguk
    Cho, Hyunsouk
    IEEE ACCESS, 2024, 12 : 14324 - 14333
  • [5] Deep multimodal learning for cross-modal retrieval: One model for all tasks
    Beltran, L. Viviana Beltran
    Caicedo, Juan C.
    Journet, Nicholas
    Coustaty, Mickael
    Lecellier, Francois
    Doucet, Antoine
    PATTERN RECOGNITION LETTERS, 2021, 146 : 38 - 45
  • [6] Meta-learning of Text Classification Tasks
    Madrid, Jorge G.
    Jair Escalante, Hugo
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS (CIARP 2019), 2019, 11896 : 107 - 119
  • [7] Crowdsourcing with Meta-Knowledge Transfer (Student Abstract)
    Xu, Sunyue
    Zhang, Jing
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 13095 - 13096
  • [8] Meta-Knowledge Transfer for Inductive Knowledge Graph Embedding
    Chen, Mingyang
    Zhang, Wen
    Zhu, Yushan
    Zhou, Hongting
    Yuan, Zonggang
    Xu, Changliang
    Chen, Huajun
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 927 - 937
  • [9] Meta Self-Paced Learning for Cross-Modal Matching
    Wei, Jiwei
    Xu, Xing
    Wang, Zheng
    Wang, Guoqing
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3835 - 3843
  • [10] Meta-Learning Based Tasks Similarity Representation for Cross Domain Lifelong Learning
    Shen, Mingge
    Chen, Dehu
    Ren, Teng
    IEEE ACCESS, 2023, 11 : 36692 - 36701