Multi-modal deep learning for automated assembly of periapical radiographs

被引:2
|
作者
Pfaender, L. [1 ]
Schneider, L. [1 ,2 ]
Buettner, M. [1 ,2 ]
Krois, J. [2 ]
Meyer-Lueckel, H. [3 ]
Schwendicke, F. [1 ,2 ,4 ,5 ,6 ,7 ]
机构
[1] Charite Univ Med Berlin, Dept Oral Diagnost Digital Hlth & Hlth Serv Res, D-14197 Berlin, Germany
[2] ITU WHO Focus Grp AI4Hlth, Top Grp Dent Diagnost & Digital Dent, Geneva, Switzerland
[3] Univ Bern, Dept Restorat Prevent & Pediat Dent, zmk Bern, Bern, Switzerland
[4] Charite Univ Med Berlin, Dept Oral Diagnost Digital Hlth & Hlth Serv Res, Assmannshauser Str 4-6, D-14197 Berlin, Germany
[5] Free Univ Berlin, Assmannshauser Str 4-6, D-14197 Berlin, Germany
[6] Humboldt Univ, Assmannshauser Str 4-6, D-14197 Berlin, Germany
[7] Berlin Inst Hlth Berlin, Assmannshauser Str 4-6, D-14197 Berlin, Germany
关键词
Deep learning; Multi-modal learning; CNNs computer vision; Time series analysis; Periapical radiographs;
D O I
10.1016/j.jdent.2023.104588
中图分类号
R78 [口腔科学];
学科分类号
1003 ;
摘要
Objectives: Periapical radiographs are oftentimes taken in series to display all teeth present in the oral cavity. Our aim was to automatically assemble such a series of periapical radiographs into an anatomically correct status using a multi-modal deep learning model.Methods: 4,707 periapical images from 387 patients (on average, 12 images per patient) were used. Radiographs were labeled according to their field of view and the dataset split into a training, validation, and test set, stratified by patient. In addition to the radiograph the timestamp of image generation was extracted and abstracted as follows: A matrix, containing the normalized timestamps of all images of a patient was constructed, representing the order in which images were taken, providing temporal context information to the deep learning model. Using the image data together with the time sequence data a multi-modal deep learning model consisting of two residual convolutional neural networks (ResNet-152 for image data, ResNet-50 for time data) was trained. Additionally, two uni-modal models were trained on image data and time data, respectively. A custom scoring technique was used to measure model performance.Results: Multi-modal deep learning outperformed both uni-modal image-based learning (p<0.001) and time-based learning (p<0.05). The multi-modal deep learning model predicted tooth labels with an F1-score, sensi-tivity and precision of 0.79, respectively, and an accuracy of 0.99. 37 out of 77 patient datasets were fully correctly assembled by multi-modal learning; in the remaining ones, usually only one image was incorrectly labeled.Conclusions: Multi-modal modeling allowed automated assembly of periapical radiographs and outperformed both uni-modal models. Dental machine learning models can benefit from additional data modalities.Clinical significance: Like humans, deep learning models may profit from multiple data sources for decision-making. We demonstrate how multi-modal learning can assist assembling periapical radiographs into an anatomically correct status. Multi-modal learning should be considered for more complex tasks, as clinically a wealth of data is usually available and could be leveraged.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Deep Multi-modal Latent Representation Learning for Automated Dementia Diagnosis
    Zhou, Tao
    Liu, Mingxia
    Fu, Huazhu
    Wang, Jun
    Shen, Jianbing
    Shao, Ling
    Shen, Dinggang
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT IV, 2019, 11767 : 629 - 638
  • [2] Automated feature detection in dental periapical radiographs by using deep learning
    Khan, Hassan Aqeel
    Haider, Muhammad Ali
    Ansari, Hassan Ali
    Ishaq, Hamna
    Kiyani, Amber
    Sohail, Kanwal
    Muhammad, Muhammad
    Khurram, Syed Ali
    ORAL SURGERY ORAL MEDICINE ORAL PATHOLOGY ORAL RADIOLOGY, 2021, 131 (06): : 711 - 720
  • [3] Multi-modal deep learning for landform recognition
    Du, Lin
    You, Xiong
    Li, Ke
    Meng, Liqiu
    Cheng, Gong
    Xiong, Liyang
    Wang, Guangxia
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2019, 158 : 63 - 75
  • [4] Deep Multi-modal Learning with Cascade Consensus
    Yang, Yang
    Wu, Yi-Feng
    Zhan, De-Chuan
    Jiang, Yuan
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2018, 11013 : 64 - 72
  • [5] Multi-modal deep distance metric learning
    Roostaiyan, Seyed Mahdi
    Imani, Ehsan
    Baghshah, Mahdieh Soleymani
    INTELLIGENT DATA ANALYSIS, 2017, 21 (06) : 1351 - 1369
  • [6] Multi-Modal Song Mood Detection with Deep Learning
    Pyrovolakis, Konstantinos
    Tzouveli, Paraskevi
    Stamou, Giorgos
    SENSORS, 2022, 22 (03)
  • [7] A Multi-Modal Deep Learning Approach for Emotion Recognition
    Shahzad, H. M.
    Bhatti, Sohail Masood
    Jaffar, Arfan
    Rashid, Muhammad
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (02): : 1561 - 1570
  • [8] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    INFORMATION FUSION, 2021, 67 : 136 - 146
  • [9] Deep contrastive representation learning for multi-modal clustering
    Lu, Yang
    Li, Qin
    Zhang, Xiangdong
    Gao, Quanxue
    NEUROCOMPUTING, 2024, 581
  • [10] Deep Learning-Based Detection of Periapical Lesions in Periapical Radiographs
    Liu, Jian
    Hong, Yanqing
    Shao, Yu
    Gao, Yongzhen
    Pan, Kexu
    Jin, Chaoran
    Du, Yi
    Yu, Xijiao
    JOURNAL OF MEDICAL AND BIOLOGICAL ENGINEERING, 2024, 44 (05) : 676 - 684