Task-aware asynchronous multi-task model with class incremental contrastive learning for surgical scene understanding

被引:0
|
作者
Seenivasan, Lalithkumar [1 ]
Islam, Mobarakol [2 ]
Xu, Mengya [1 ]
Lim, Chwee Ming [3 ]
Ren, Hongliang [1 ,4 ,5 ]
机构
[1] Natl Univ Singapore, Dept Biomed Engn, Singapore, Singapore
[2] Imperial Coll London, Dept Comp, London, England
[3] Singapore Gen Hosp, Head & Neck Surg, Singapore, Singapore
[4] Chinese Univ Hong Kong, Dept Elect Engn, Shatin, Hong Kong, Peoples R China
[5] Chinese Univ Hong Kong, Shun Hing Inst Adv Engn, Shatin, Hong Kong, Peoples R China
基金
国家重点研发计划;
关键词
Surgical scene understanding; Domain generalization; Scene graph; Curriculum learning;
D O I
10.1007/s11548-022-02800-2
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
PurposeSurgery scene understanding with tool-tissue interaction recognition and automatic report generation can play an important role in intra-operative guidance, decision-making and postoperative analysis in robotic surgery. However, domain shifts between different surgeries with inter and intra-patient variation and novel instruments' appearance degrade the performance of model prediction. Moreover, it requires output from multiple models, which can be computationally expensive and affect real-time performance.MethodologyA multi-task learning (MTL) model is proposed for surgical report generation and tool-tissue interaction prediction that deals with domain shift problems. The model forms of shared feature extractor, mesh-transformer branch for captioning and graph attention branch for tool-tissue interaction prediction. The shared feature extractor employs class incremental contrastive learning to tackle intensity shift and novel class appearance in the target domain. We design Laplacian of Gaussian-based curriculum learning into both shared and task-specific branches to enhance model learning. We incorporate a task-aware asynchronous MTL optimization technique to fine-tune the shared weights and converge both tasks optimally.ResultsThe proposed MTL model trained using task-aware optimization and fine-tuning techniques reported a balanced performance (BLEU score of 0.4049 for scene captioning and accuracy of 0.3508 for interaction detection) for both tasks on the target domain and performed on-par with single-task models in domain adaptation.ConclusionThe proposed multi-task model was able to adapt to domain shifts, incorporate novel instruments in the target domain, and perform tool-tissue interaction detection and report generation on par with single-task models.
引用
收藏
页码:921 / 928
页数:8
相关论文
共 50 条
  • [1] Task-aware asynchronous multi-task model with class incremental contrastive learning for surgical scene understanding
    Lalithkumar Seenivasan
    Mobarakol Islam
    Mengya Xu
    Chwee Ming Lim
    Hongliang Ren
    International Journal of Computer Assisted Radiology and Surgery, 2023, 18 : 921 - 928
  • [2] Task-Aware Dynamic Model Optimization for Multi-Task Learning
    Choi, Sujin
    Jin, Hyundong
    Kim, Eunwoo
    IEEE ACCESS, 2023, 11 : 137709 - 137717
  • [3] AdaTask: A Task-Aware Adaptive Learning Rate Approach to Multi-Task Learning
    Yang, Enneng
    Pan, Junwei
    Wang, Ximei
    Yu, Haibin
    Shen, Li
    Chen, Xihua
    Xiao, Lei
    Jiang, Jie
    Guo, Guibing
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10745 - 10753
  • [4] Global-Reasoned Multi-Task Learning Model for Surgical Scene Understanding
    Seenivasan, Lalithkumar
    Mitheran, Sai
    Islam, Mobarakol
    Ren, Hongliang
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 3858 - 3865
  • [5] Asynchronous Multi-Task Learning
    Baytas, Inci M.
    Yan, Ming
    Jain, Anil K.
    Zhou, Jiayu
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2016, : 11 - 20
  • [6] Multi-task Graph Neural Architecture Search with Task-aware Collaboration and Curriculum
    Qin, Yijian
    Wang, Xin
    Zhang, Ziwei
    Chen, Hong
    Zhu, Wenwu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] HirMTL: Hierarchical Multi-Task Learning for dense scene understanding
    Luo, Huilan
    Hu, Weixia
    Wei, Yixiao
    He, Jianlong
    Yu, Minghao
    NEURAL NETWORKS, 2025, 181
  • [8] Episodic task agnostic contrastive training for multi-task learning?
    Zhou, Fan
    Chen, Yuyi
    Wen, Jun
    Zeng, Qiuhao
    Shui, Changjian
    Ling, Charles X.
    Yang, Shichun
    Wang, Boyu
    NEURAL NETWORKS, 2023, 162 : 34 - 45
  • [9] Learning Task-Aware Language-Image Representation for Class-Incremental Object Detection
    Zhang, Hongquan
    Gao, Bin-Bin
    Zeng, Yi
    Tian, Xudong
    Tan, Xin
    Zhang, Zhizhong
    Qu, Yanyun
    Liu, Jun
    Xie, Yuan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 7096 - 7104
  • [10] Multi-Task Learning Model for Kazakh Query Understanding
    Haisa, Gulizada
    Altenbek, Gulila
    SENSORS, 2022, 22 (24)