Multimodal registration across 3D point clouds and CT-volumes

被引:9
作者
Saiti E. [1 ]
Theoharis T. [1 ]
机构
[1] Norwegian University of Science and Technology (NTNU), Department of Computer and Information Science
来源
Computers and Graphics (Pergamon) | 2022年 / 106卷
基金
欧盟地平线“2020”;
关键词
3D point cloud; 3D registration; 3D volume; Alignment; Data fusion; Multimodal;
D O I
10.1016/j.cag.2022.06.012
中图分类号
学科分类号
摘要
Multimodal registration is a challenging problem in visual computing, commonly faced during medical image-guided interventions, data fusion and 3D object retrieval. The main challenge of multimodal registration is finding accurate correspondence between modalities, since different modalities do not exhibit the same characteristics. This paper explores how the coherence of different modalities can be utilized for the challenging task of 3D multimodal registration. A novel deep learning multimodal registration framework is proposed by introducing a siamese deep learning architecture, especially designed for aligning and fusing modalities of different structural and physical principles. The cross-modal attention blocks lead the network to establish correspondences between features of different modalities. The proposed framework focuses on the alignment of 3D point clouds and the micro-CT 3D volumes of the same object. A multimodal dataset consisting of real micro-CT scans and their synthetically generated 3D models (point clouds) is presented and utilized for evaluating our methodology. © 2022 The Author(s)
引用
收藏
页码:259 / 266
页数:7
相关论文
共 63 条
[1]  
Liu A.-A., Guo F.-B., Zhou H.-Y., Yan C.-G., Gao Z., Li X.-Y., Li W.-H., Domain-adversarial-guided siamese network for unsupervised cross-domain 3-D object retrieval, IEEE Trans Cybern, (2022)
[2]  
Chen Z., Jing L., Liang Y., Tian Y., Li B., Multimodal semi-supervised learning for 3D objects, (2021)
[3]  
Joaquim M.J.F.A.F., Jorge A., Towards 3D modeling using sketches and retrieval, Eurographics workshop on sketch-based interfaces and modeling 2004, (2004)
[4]  
Ruan Y., Lee H.-H., Zhang K., Chang A.X., TriCoLo: TRimodal contrastive loss for fine-grained text to shape retrieval, (2022)
[5]  
Yin T., Zhou X., Krahenbuhl P., Multimodal virtual point 3D detection, Adv Neural Inf Process Syst, 34, (2021)
[6]  
Hess M., Petrovic V., Meyer D., Rissolo D., Kuester F., Fusion of multimodal three-dimensional data for comprehensive digital documentation of cultural heritage sites, 2015 digital heritage, Vol. 2, pp. 595-602, (2015)
[7]  
Hervella A.S., Rouco J., Novo J., Ortega M., Multimodal registration of retinal images using domain-specific landmarks and vessel enhancement, Procedia Comput Sci, 126, pp. 97-104, (2018)
[8]  
Zhan K., Fritsch D., Wagner J., Integration of photogrammetry, computed tomography and endoscopy for gyroscope 3D digitization, Int Arch Photogramm Remote Sens Spat Inf Sci, 46, pp. 925-931, (2021)
[9]  
Ramos M.M., Remondino F., Data fusion in cultural heritage-A review, Int Arch Photogramm Remote Sens Spat Inf Sci, 40, 5, (2015)
[10]  
Mannes D., Schmid F., Frey J., Schmidt-Ott K., Lehmann E., Combined neutron and X-ray imaging for non-invasive investigations of cultural heritage objects, Physics Procedia, 69, pp. 653-660, (2015)