A multimodal transformer to fuse images and metadata for skin disease classification

被引:0
|
作者
Gan Cai
Yu Zhu
Yue Wu
Xiaoben Jiang
Jiongyao Ye
Dawei Yang
机构
[1] East China University of Science and Technology,School of Information Science and Engineering
[2] Zhongshan Hospital,Department of Pulmonary and Critical Care Medicine
[3] Fudan University,undefined
[4] Shanghai Engineering Research Center of Internet of Things for Respiratory Medicine,undefined
来源
The Visual Computer | 2023年 / 39卷
关键词
Skin disease; Deep learning; Transformer; Multimodal fusion; Attention;
D O I
暂无
中图分类号
学科分类号
摘要
Skin disease cases are rising in prevalence, and the diagnosis of skin diseases is always a challenging task in the clinic. Utilizing deep learning to diagnose skin diseases could help to meet these challenges. In this study, a novel neural network is proposed for the classification of skin diseases. Since the datasets for the research consist of skin disease images and clinical metadata, we propose a novel multimodal Transformer, which consists of two encoders for both images and metadata and one decoder to fuse the multimodal information. In the proposed network, a suitable Vision Transformer (ViT) model is utilized as the backbone to extract image deep features. As for metadata, they are regarded as labels and a new Soft Label Encoder (SLE) is designed to embed them. Furthermore, in the decoder part, a novel Mutual Attention (MA) block is proposed to better fuse image features and metadata features. To evaluate the model’s effectiveness, extensive experiments have been conducted on the private skin disease dataset and the benchmark dataset ISIC 2018. Compared with state-of-the-art methods, the proposed model shows better performance and represents an advancement in skin disease diagnosis.
引用
收藏
页码:2781 / 2793
页数:12
相关论文
共 50 条
  • [21] Adversarial transformer network for classification of lung cancer disease from CT scan images
    Murthy, S. V. S. N.
    Prasad, P. Murali Krishna
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 86
  • [22] Multimodal diagnosis model of Alzheimer's disease based on improved Transformer
    Tang, Yan
    Xiong, Xing
    Tong, Gan
    Yang, Yuan
    Zhang, Hao
    BIOMEDICAL ENGINEERING ONLINE, 2024, 23 (01)
  • [23] A Deep CNN Transformer Hybrid Model for Skin Lesion Classification of Dermoscopic Images Using Focal Loss
    Nie, Yali
    Sommella, Paolo
    Carratu, Marco
    O'Nils, Mattias
    Lundgren, Jan
    DIAGNOSTICS, 2023, 13 (01)
  • [24] Multimodal diagnosis model of Alzheimer’s disease based on improved Transformer
    Yan Tang
    Xing Xiong
    Gan Tong
    Yuan Yang
    Hao Zhang
    BioMedical Engineering OnLine, 23
  • [25] Research on Multimodal Sentiment Classification of Internet Memes Based on Transformer
    Chi, Shengnan
    Sang, Guoming
    Shi, Xian
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CRYPTOGRAPHY, NETWORK SECURITY AND COMMUNICATION TECHNOLOGY, CNSCT 2024, 2024, : 445 - 450
  • [26] Transforming Alzheimer's Disease Diagnosis: Implementing Vision Transformer (ViT) for MRI Images Classification
    Kurniasari, Dian
    Pratama, Muhammad Dwi
    Junaidi, Akmal
    Faisol, Ahmad
    JOURNAL OF INFORMATION AND COMMUNICATION TECHNOLOGY-MALAYSIA, 2025, 24 (01): : 130 - 152
  • [27] Stellar Classification with Vision Transformer and SDSS Photometric Images
    Yang, Yi
    Li, Xin
    UNIVERSE, 2024, 10 (05)
  • [28] A transformer-based unified multimodal framework for Alzheimer's disease assessment
    Department of Big Data in Health Science, School of Public Health and Center of Clinical Big Data and Analytics of The Second Affiliated Hospital, Zhejiang University School of Medicine, Zhejiang, Hangzhou, China
    不详
    130024, China
    Comput. Biol. Med.,
  • [29] Transformer based fruits disease classification
    Zala S.
    Goyal V.
    Sharma S.
    Shukla A.
    Multimedia Tools and Applications, 2025, 84 (4) : 1627 - 1647
  • [30] Multimodal skin lesion classification using deep learning
    Yap, Jordan
    Yolland, William
    Tschandl, Philipp
    EXPERIMENTAL DERMATOLOGY, 2018, 27 (11) : 1261 - 1267