Medical Publication Recommendation Based on Cross-Modal Contrastive Learning Between Knowledge and Graph

被引:0
作者
Xia, Zhonghua [1 ]
Qi, Jianglei [2 ]
Ding, Hao [3 ,4 ]
机构
[1] School of Information Management, Nanjing University, Nanjing
[2] School of Civil Affairs and Social Work, Changsha Social Work College, Changsha
[3] College of Education Science and Technology, Nanjing University of Posts and Telecommunications, Nanjing
[4] School of Management, Nanjing University of Posts and Telecommunications, Nanjing
基金
中国国家社会科学基金;
关键词
Contrastive Learning; Cross Attention; Medical Publication Recommendations; Recommender Systems;
D O I
10.11925/infotech.2096-3467.2024.0022
中图分类号
学科分类号
摘要
[Objective] This study proposes a medical publication recommendation model that uses cross-modal information to improve recommendation accuracy. [Methods] First, the medical terminology system was employed to standardize label content and align image-text tags. Paired semantic labels were then utilized to align feature semantics between images and texts through contrastive learning. Based on the aligned semantic features, a cross-modal cross-attention mechanism was constructed, and user preferences for publications were predicted by analyzing their interest weights across different modalities. [Results] Comparative experiments with three state-of-the-art multimodal baseline models on two publication datasets showed that the proposed model achieved an average precision of 62.79%, F1-score of 53.62%, and NDCG of 61.17%, outperforming the baseline models in all metrics. [Limitations] Additional cold-start methods may be required for pre-training data containing only single-modality information. [Conclusions] The proposed model exhibits strong cross-modal feature fusion capabilities, effectively mitigating semantic gaps between modalities and improving the accuracy of medical publication recommendations. © 2025 Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:136 / 145
页数:9
相关论文
共 32 条
[1]  
Kim S, Lee N, Lee J, Et al., Heterogeneous Graph Learning for Multi-modal Medical Data Analysis, Proceedings of the 37th AAAI Conference on Artificial Intelligence, pp. 5141-5150, (2023)
[2]  
Liu H, Li S S, Zhu J C, Et al., DDIFN: A Dual-discriminator Multi-modal Medical Image Fusion Network, ACM Transactions on Multimedia Computing, Communications and Applications, 19, 4, (2023)
[3]  
Ding Hao, Ai Wenhua, Hu Guangwei, Et al., A Personalized Recommendation Model with Time Series Fluctuation of User Interest, Data Analysis and Knowledge Discovery, 5, 11, pp. 45-58, (2021)
[4]  
Song Huanying, Jia Xuewen, Research Progress, Themes, and Characteristics of Digital Reading in China, Chinese Editor, 1, pp. 78-84, (2022)
[5]  
de Souza M A, Cordeiro D C A, de Oliveira J, Et al., 3D Multi-modality Medical Imaging: Combining Anatomical and Infrared Thermal Images for 3D Reconstruction, Sensors, 23, 3, (2023)
[6]  
Zhang S, Zhang J J, Tian B, Et al., Multi-modal Contrastive Mutual Learning and Pseudo-Label Re-learning for Semi-supervised Medical Image Segmentation, Medical Image Analysis, 83, (2023)
[7]  
Abid M H, Ashraf R, Mahmood T, Et al., Multi-modal Medical Image Classification Using Deep Residual Network and Genetic Algorithm, PLoS One, 18, 6, (2023)
[8]  
Vasu G T, Palanisamy P., CT and MRI Multi-modal Medical Image Fusion Using Weight-Optimized Anisotropic Diffusion Filtering, Soft Computing, 27, 13, pp. 9105-9117, (2023)
[9]  
Wang B Y, Xie Q Q, Pei J H, Et al., Pre-trained Language Models in Biomedical Domain: A Systematic Survey, ACM Computing Surveys, 56, 3, (2023)
[10]  
Dogra A, Kumar S., Multi-modality Medical Image Fusion Based on Guided Filter and Image Statistics in Multidirectional Shearlet Transform Domain, Journal of Ambient Intelligence and Humanized Computing, 14, 9, pp. 12191-12205, (2023)