Flexible Fusion Network for Multi-Modal Brain Tumor Segmentation

被引:45
作者
Yang, Hengyi [1 ]
Zhou, Tao [1 ]
Zhou, Yi [2 ]
Zhang, Yizhe [1 ]
Fu, Huazhu [3 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, PCA Lab, Key Lab Intelligent Percept & Syst High Dimens In, Nanjing 210094, Peoples R China
[2] Southeast Univ, Sch Comp Sci & Engn, Nanjing 211189, Peoples R China
[3] ASTAR, Inst High Performance Comp, Singapore 138632, Singapore
基金
中国国家自然科学基金;
关键词
Brain tumor segmentation; magnetic resonance imaging; cross-modal feature-enhanced module; multi-modal collaboration module; MODEL;
D O I
10.1109/JBHI.2023.3271808
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automated brain tumor segmentation is crucial for aiding brain disease diagnosis and evaluating disease progress. Currently, magnetic resonance imaging (MRI) is a routinely adopted approach in the field of brain tumor segmentation that can provide different modality images. It is critical to leverage multi-modal images to boost brain tumor segmentation performance. Existing works commonly concentrate on generating a shared representation by fusing multi-modal data, while few methods take into account modality-specific characteristics. Besides, how to efficiently fuse arbitrary numbers of modalities is still a difficult task. In this study, we present a flexible fusion network (termed F(2)Net) for multi-modal brain tumor segmentation, which can flexibly fuse arbitrary numbers of multi-modal information to explore complementary information while maintaining the specific characteristics of each modality. Our F(2)Net is based on the encoderdecoder structure, which utilizes two Transformer-based feature learning streams and a cross-modal shared learning network to extract individual and shared feature representations. To effectively integrate the knowledge from the multi-modality data, we propose a cross-modal featureenhanced module (CFM) and a multi-modal collaboration module (MCM), which aims at fusing the multi-modal features into the shared learning network and incorporating the features from encoders into the shared decoder, respectively. Extensive experimental results on multiple benchmark datasets demonstrate the effectiveness of our F(2)Net over other state-of-the-art segmentation methods.
引用
收藏
页码:3349 / 3359
页数:11
相关论文
共 55 条
[1]  
Cao Hu, 2023, Computer Vision - ECCV 2022 Workshops: Proceedings. Lecture Notes in Computer Science (13803), P205, DOI 10.1007/978-3-031-25066-8_9
[2]   Learning With Privileged Multimodal Knowledge for Unimodal Segmentation [J].
Chen, Cheng ;
Dou, Qi ;
Jin, Yueming ;
Liu, Quande ;
Heng, Pheng Ann .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (03) :621-632
[3]   Robust Multimodal Brain Tumor Segmentation via Feature Disentanglement and Gated Fusion [J].
Chen, Cheng ;
Dou, Qi ;
Jin, Yueming ;
Chen, Hao ;
Qin, Jing ;
Pheng-Ann Heng .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT III, 2019, 11766 :447-456
[4]  
Chen J., 2021, arXiv
[5]  
Chen LC, 2017, Arxiv, DOI arXiv:1706.05587
[6]   Efficient multilevel brain tumor segmentation with integrated Bayesian model classification [J].
Corso, Jason J. ;
Sharon, Eitan ;
Dube, Shishir ;
El-Saden, Suzie ;
Sinha, Usha ;
Yuille, Alan .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2008, 27 (05) :629-640
[7]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[8]   MVFusFra: A Multi-View Dynamic Fusion Framework for Multimodal Brain Tumor Segmentation [J].
Ding, Yi ;
Zheng, Wei ;
Geng, Ji ;
Qin, Zhen ;
Choo, Kim-Kwang Raymond ;
Qin, Zhiguang ;
Hou, Xiaolin .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (04) :1570-1581
[9]  
Dolz J., 2018, P INT WORKSH CHALL C, P130
[10]   HyperDense-Net: A Hyper-Densely Connected CNN for Multi-Modal Image Segmentation [J].
Dolz, Jose ;
Gopinath, Karthik ;
Yuan, Jing ;
Lombaert, Herve ;
Desrosiers, Christian ;
Ben Ayed, Ismail .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (05) :1116-1126