InClust plus : the deep generative framework with mask modules for multimodal data integration, imputation, and cross-modal generation

被引:2
|
作者
Wang, Lifei [1 ]
Nie, Rui [2 ,3 ,4 ]
Miao, Xuexia [2 ,3 ]
Cai, Yankai [5 ]
Wang, Anqi [1 ]
Zhang, Hanwen [1 ]
Zhang, Jiang [6 ]
Cai, Jun [2 ,3 ,4 ]
机构
[1] Zhejiang Shuren Univ, Shulan Hangzhou Hosp, Shulan Int Med Coll, Hangzhou, Peoples R China
[2] China Natl Ctr Bioinformat, Beijing, Peoples R China
[3] Chinese Acad Sci, Beijing Inst Genom, Key Lab Genom & Precis Med, Beijing 100101, Peoples R China
[4] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[5] China Univ Geosci, Sch Econ & Management, Wuhan, Peoples R China
[6] Beijing Normal Univ, Sch Syst Sci, Beijing 100875, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep generative framework with mask modules; Multi-omics; Data integration; Cross-modal imputation; Cross-modal generation; SINGLE; OMICS; CHROMATIN; RNA;
D O I
10.1186/s12859-024-05656-2
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
BackgroundWith the development of single-cell technology, many cell traits can be measured. Furthermore, the multi-omics profiling technology could jointly measure two or more traits in a single cell simultaneously. In order to process the various data accumulated rapidly, computational methods for multimodal data integration are needed.ResultsHere, we present inClust+, a deep generative framework for the multi-omics. It's built on previous inClust that is specific for transcriptome data, and augmented with two mask modules designed for multimodal data processing: an input-mask module in front of the encoder and an output-mask module behind the decoder. InClust+ was first used to integrate scRNA-seq and MERFISH data from similar cell populations, and to impute MERFISH data based on scRNA-seq data. Then, inClust+ was shown to have the capability to integrate the multimodal data (e.g. tri-modal data with gene expression, chromatin accessibility and protein abundance) with batch effect. Finally, inClust+ was used to integrate an unlabeled monomodal scRNA-seq dataset and two labeled multimodal CITE-seq datasets, transfer labels from CITE-seq datasets to scRNA-seq dataset, and generate the missing modality of protein abundance in monomodal scRNA-seq data. In the above examples, the performance of inClust+ is better than or comparable to the most recent tools in the corresponding task.ConclusionsThe inClust+ is a suitable framework for handling multimodal data. Meanwhile, the successful implementation of mask in inClust+ means that it can be applied to other deep learning methods with similar encoder-decoder architecture to broaden the application scope of these models.
引用
收藏
页数:17
相关论文
共 5 条
  • [1] InClust+: the deep generative framework with mask modules for multimodal data integration, imputation, and cross-modal generation
    Lifei Wang
    Rui Nie
    Xuexia Miao
    Yankai Cai
    Anqi Wang
    Hanwen Zhang
    Jiang Zhang
    Jun Cai
    BMC Bioinformatics, 25
  • [2] Audio-Visual Cross-Modal Generation with Multimodal Variational Generative Model
    Xu, Zhubin
    Wang, Tianlei
    Liu, Dekang
    Hu, Dinghan
    Zeng, Huangiang
    Cao, Jiuwen
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [3] Deep Generative Cross-modal On-body Accelerometer Data Synthesis from Videos
    Zhang, Shibo
    Alshurafa, Nabil
    UBICOMP/ISWC '20 ADJUNCT: PROCEEDINGS OF THE 2020 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING AND PROCEEDINGS OF THE 2020 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS, 2020, : 223 - 227
  • [4] CRISP: A cross-modal integration framework based on the surprisingly popular algorithm for multimodal named entity recognition
    Liu, Haitao
    Xin, Xianwei
    Song, Jihua
    Peng, Weiming
    NEUROCOMPUTING, 2025, 614
  • [5] scCross: a deep generative model for unifying single-cell multi-omics with seamless integration, cross-modal generation, and in silico exploration
    Yang, Xiuhui
    Mann, Koren K.
    Wu, Hao
    Ding, Jun
    GENOME BIOLOGY, 2024, 25 (01):