Deep Low-Shot Learning for Biological Image Classification and Visualization From Limited Training Samples

被引:4
作者
Cai, Lei [1 ]
Wang, Zhengyang [2 ]
Kulathinal, Rob [3 ]
Kumar, Sudhir [3 ]
Ji, Shuiwang [2 ]
机构
[1] Washington State Univ, Sch Elect Engn & Comp Sci, Pullman, WA 99164 USA
[2] Texas A&M Univ, Dept Comp Sci & Engn, College Stn, TX 77843 USA
[3] Temple Univ, Dept Biol, Philadelphia, PA 19122 USA
基金
美国国家科学基金会;
关键词
Training; Task analysis; Feature extraction; Biological system modeling; Deep learning; Annotations; Biological information theory; Biological image classification; deep two-step low-shot learning; Drosophila in situ hybridization (ISH) images; limited training samples; model interpretation and visualization; GENE-EXPRESSION PATTERNS; AUTOMATED ANNOTATION; DROSOPHILA;
D O I
10.1109/TNNLS.2021.3106831
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Predictive modeling is useful but very challenging in biological image analysis due to the high cost of obtaining and labeling training data. For example, in the study of gene interaction and regulation in Drosophila embryogenesis, the analysis is most biologically meaningful when in situ hybridization (ISH) gene expression pattern images from the same developmental stage are compared. However, labeling training data with precise stages is very time-consuming even for developmental biologists. Thus, a critical challenge is how to build accurate computational models for precise developmental stage classification from limited training samples. In addition, identification and visualization of developmental landmarks are required to enable biologists to interpret prediction results and calibrate models. To address these challenges, we propose a deep two-step low-shot learning framework to accurately classify ISH images using limited training images. Specifically, to enable accurate model training on limited training samples, we formulate the task as a deep low-shot learning problem and develop a novel two-step learning approach, including data-level learning and feature-level learning. We use a deep residual network as our base model and achieve improved performance in the precise stage prediction task of ISH images. Furthermore, the deep model can be interpreted by computing saliency maps, which consists of pixel-wise contributions of an image to its prediction result. In our task, saliency maps are used to assist the identification and visualization of developmental landmarks. Our experimental results show that the proposed model can not only make accurate predictions but also yield biologically meaningful interpretations. We anticipate our methods to be easily generalizable to other biological image classification tasks with small training datasets. Our open-source code is available at https://github.com/divelab/lsl-fly.
引用
收藏
页码:2528 / 2538
页数:11
相关论文
共 44 条
  • [11] Identity Mappings in Deep Residual Networks
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 : 630 - 645
  • [12] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [13] Automated annotation of Drosophila gene expression patterns using a controlled vocabulary
    Ji, Shuiwang
    Sun, Liang
    Jin, Rong
    Kumar, Sudhir
    Ye, Jieping
    [J]. BIOINFORMATICS, 2008, 24 (17) : 1881 - 1888
  • [14] Ji SW, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P407
  • [15] A bag-of-words approach for Drosophila gene expression pattern annotation
    Ji, Shuiwang
    Li, Ying-Xin
    Zhou, Zhi-Hua
    Kumar, Sudhir
    Ye, Jieping
    [J]. BMC BIOINFORMATICS, 2009, 10
  • [16] Comparison of embryonic expression within multigene families using the flyexpress discovery platform reveals more spatial than temporal divergence
    Konikoff, Charlotte E.
    Karr, Timothy L.
    McCutchan, Michael
    Newfeld, Stuart J.
    Kumar, Sudhir
    [J]. DEVELOPMENTAL DYNAMICS, 2012, 241 (01) : 150 - 160
  • [17] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90
  • [18] Kumar S, 2002, GENETICS, V162, P2037
  • [19] FlyExpress: visual mining of spatiotemporal patterns for genes and publications in Drosophila embryogenesis
    Kumar, Sudhir
    Konikoff, Charlotte
    Van Emden, Bernard
    Busick, Christopher
    Davis, Kailah T.
    Ji, Shuiwang
    Wu, Lin-Wei
    Ramos, Hector
    Brody, Thomas
    Panchanathan, Sethuraman
    Ye, Jieping
    Karr, Timothy L.
    Gerold, Kristyn
    McCutchan, Michael
    Newfeld, Stuart J.
    [J]. BIOINFORMATICS, 2011, 27 (23) : 3319 - 3320
  • [20] Computational identification of Drosophila microRNA genes -: art. no. R42
    Lai, EC
    Tomancak, P
    Williams, RW
    Rubin, GM
    [J]. GENOME BIOLOGY, 2003, 4 (07)