Wireless capsule endoscopy multiclass classification using three-dimensional deep convolutional neural network model

被引:1
作者
Bordbar, Mehrdokht [1 ]
Helfroush, Mohammad Sadegh [1 ]
Danyali, Habibollah [1 ]
Ejtehadi, Fardad [2 ]
机构
[1] Shiraz Univ Technol, Dept Elect Engn, Shiraz, Iran
[2] Shiraz Univ Med Sci, Gastroenterohepatol Res Ctr, Sch Med, Dept Internal Med, Shiraz, Iran
关键词
Wireless capsule endoscopy; Image classification; Deep learning; 3D convolutional neural network; IMAGES;
D O I
10.1186/s12938-023-01186-9
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
BackgroundWireless capsule endoscopy (WCE) is a patient-friendly and non-invasive technology that scans the whole of the gastrointestinal tract, including difficult-to-access regions like the small bowel. Major drawback of this technology is that the visual inspection of a large number of video frames produced during each examination makes the physician diagnosis process tedious and prone to error. Several computer-aided diagnosis (CAD) systems, such as deep network models, have been developed for the automatic recognition of abnormalities in WCE frames. Nevertheless, most of these studies have only focused on spatial information within individual WCE frames, missing the crucial temporal data within consecutive frames.MethodsIn this article, an automatic multiclass classification system based on a three-dimensional deep convolutional neural network (3D-CNN) is proposed, which utilizes the spatiotemporal information to facilitate the WCE diagnosis process. The 3D-CNN model fed with a series of sequential WCE frames in contrast to the two-dimensional (2D) model, which exploits frames as independent ones. Moreover, the proposed 3D deep model is compared with some pre-trained networks. The proposed models are trained and evaluated with 29 subject WCE videos (14,691 frames before augmentation). The performance advantages of 3D-CNN over 2D-CNN and pre-trained networks are verified in terms of sensitivity, specificity, and accuracy.Results3D-CNN outperforms the 2D technique in all evaluation metrics (sensitivity: 98.92 vs. 98.05, specificity: 99.50 vs. 86.94, accuracy: 99.20 vs. 92.60). In conclusion, a novel 3D-CNN model for lesion detection in WCE frames is proposed in this study.ConclusionThe results indicate the performance of 3D-CNN over 2D-CNN and some well-known pre-trained classifier networks. The proposed 3D-CNN model uses the rich temporal information in adjacent frames as well as spatial data to develop an accurate and efficient model.
引用
收藏
页数:23
相关论文
共 70 条
  • [1] Adewole S, P FUTURE TECHNOLOGIE, V2
  • [2] Application of Convolutional Neural Networks for Automated Ulcer Detection in Wireless Capsule Endoscopy Images
    Alaskar, Haya
    Hussain, Abir
    Al-Aseem, Nourah
    Liatsis, Panos
    Al-Jumeily, Dhiya
    [J]. SENSORS, 2019, 19 (06)
  • [3] Classification and Visualisation of Normal and Abnormal Radiographs; A Comparison between Eleven Convolutional Neural Network Architectures
    Ananda, Ananda
    Ngan, Kwun Ho
    Karabag, Cefa
    Ter-Sarkisov, Aram
    Alonso, Eduardo
    Reyes-Aldasoro, Constantino Carlos
    [J]. SENSORS, 2021, 21 (16)
  • [4] Automatic detection of erosions and ulcerations in wireless capsule endoscopy images based on a deep convolutional neural network
    Aoki, Tomonori
    Yamada, Atsuo
    Aoyama, Kazuharu
    Saito, Hiroaki
    Tsuboi, Akiyoshi
    Nakada, Ayako
    Niikura, Ryota
    Fujishiro, Mitsuhiro
    Oka, Shiro
    Ishihara, Soichiro
    Matsuda, Tomoki
    Tanaka, Shinji
    Koike, Kazuhiko
    Tada, Tomohiro
    [J]. GASTROINTESTINAL ENDOSCOPY, 2019, 89 (02) : 357 - +
  • [5] A comparative study of fourteen deep learning networks for multi skin lesion classification (MSLC) on unbalanced data
    Arora, Ginni
    Dubey, Ashwani Kumar
    Jaffery, Zainul Abdin
    Rocha, Alvaro
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (11) : 7989 - 8015
  • [6] ViBe: A Universal Background Subtraction Algorithm for Video Sequences
    Barnich, Olivier
    Van Droogenbroeck, Marc
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2011, 20 (06) : 1709 - 1724
  • [7] Ozyoruk KB, 2020, Arxiv, DOI [arXiv:2006.16670, DOI 10.1016/J.MEDIA.2021.102058, 10.1016/j.media.2021.102058]
  • [8] Position statement on priorities for artificial intelligence in GI endoscopy: a report by the ASGE Task Force
    Berzin, Tyler M.
    Parasa, Sravanthi
    Wallace, Michael B.
    Gross, Seth A.
    Repici, Alessandro
    Sharma, Prateek
    [J]. GASTROINTESTINAL ENDOSCOPY, 2020, 92 (04) : 951 - 959
  • [9] Improving Temporal Stability and Accuracy for Endoscopic Video Tissue Classification Using Recurrent Neural Networks
    Boers, Tim
    van der Putten, Joost
    Struyvenberg, Maarten
    Fockens, Kiki
    Jukema, Jelmer
    Schoon, Erik
    van der Sommen, Fons
    Bergman, Jacques
    de With, Peter
    [J]. SENSORS, 2020, 20 (15) : 1 - 11
  • [10] Real-time differentiation of adenomatous and hyperplastic diminutive colorectal polyps during analysis of unaltered videos of standard colonoscopy using a deep learning model
    Byrne, Michael F.
    Chapados, Nicolas
    Soudan, Florian
    Oertel, Clemens
    Linares Perez, Milagros
    Kelly, Raymond
    Iqbal, Nadeem
    Chandelier, Florent
    Rex, Douglas K.
    [J]. GUT, 2019, 68 (01) : 94 - 100