Object and anatomical feature recognition in surgical video images based on a convolutional neural network

被引:18
作者
Bamba, Yoshiko [1 ]
Ogawa, Shimpei [1 ]
Itabashi, Michio [1 ]
Shindo, Hironari [2 ]
Kameoka, Shingo [3 ]
Okamoto, Takahiro [4 ]
Yamamoto, Masakazu [1 ]
机构
[1] Tokyo Womens Med Univ, Inst Gastroenterol, Dept Surg, Shinjuku Ku, 8-1 Kawadacho, Tokyo 1628666, Japan
[2] Otsuki Municipal Cent Hosp, Yamanashi, Japan
[3] Ushiku Aiwa Hosp, Ibaraki, Japan
[4] Tokyo Womens Med Univ, Dept Breast Endocrinol Surg, Tokyo, Japan
关键词
Image-guided navigation technology; Surgical education; Convolutional neural network; Computer vision; Object detection; GASTRIC-CANCER; SURGERY;
D O I
10.1007/s11548-021-02434-w
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Purpose Artificial intelligence-enabled techniques can process large amounts of surgical data and may be utilized for clinical decision support to recognize or forecast adverse events in an actual intraoperative scenario. To develop an image-guided navigation technology that will help in surgical education, we explored the performance of a convolutional neural network (CNN)-based computer vision system in detecting intraoperative objects. Methods The surgical videos used for annotation were recorded during surgeries conducted in the Department of Surgery of Tokyo Women's Medical University from 2019 to 2020. Abdominal endoscopic images were cut out from manually captured surgical videos. An open-source programming framework for CNN was used to design a model that could recognize and segment objects in real time through IBM Visual Insights. The model was used to detect the GI tract, blood, vessels, uterus, forceps, ports, gauze and clips in the surgical images. Results The accuracy, precision and recall of the model were 83%, 80% and 92%, respectively. The mean average precision (mAP), the calculated mean of the precision for each object, was 91%. Among surgical tools, the highest recall and precision of 96.3% and 97.9%, respectively, were achieved for forceps. Among the anatomical structures, the highest recall and precision of 92.9% and 91.3%, respectively, were achieved for the GI tract. Conclusion The proposed model could detect objects in operative images with high accuracy, highlighting the possibility of using AI-based object recognition techniques for intraoperative navigation. Real-time object recognition will play a major role in navigation surgery and surgical education.
引用
收藏
页码:2045 / 2054
页数:10
相关论文
empty
未找到相关数据