Deep Model Based Transfer and Multi-Task Learning for Biological Image Analysis

被引:0
作者
Zhang, Wenlu [1 ]
Li, Rongjian [1 ]
Zeng, Tao [1 ]
Sun, Qian [2 ]
Kumar, Sudhir [3 ,4 ,5 ]
Ye, Jieping [6 ,7 ]
Ji, Shuiwang [1 ]
机构
[1] Old Dominion Univ, Dept Comp Sci, Norfolk, VA 23529 USA
[2] Arizona State Univ, Dept Comp Sci & Engn, Tempe, AZ 85287 USA
[3] Temple Univ, Inst Genom & Evolutionary Med, Philadelphia, PA 19122 USA
[4] Temple Univ, Dept Biol, Philadelphia, PA 19122 USA
[5] King Abdulaziz Univ, Ctr Excellence Genom Med Res, Jeddah, Saudi Arabia
[6] Univ Michigan, Dept Computat Med & Bioinformat, Ann Arbor, MI 48109 USA
[7] Univ Michigan, Dept Elect Engn & Comp Sci, Ann Arbor, MI 48109 USA
来源
KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING | 2015年
关键词
Deep learning; transfer learning; multi-task learning; image analysis; bioinformatics; GENE-EXPRESSION PATTERNS; AUTOMATED ANNOTATION;
D O I
10.1145/2783258.2783304
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A central theme in learning from image data is to develop appropriate image representations for the specific task at hand. Traditional methods used handcrafted local features combined with high-level image representations to generate image-level representations. Thus, a practical challenge is to determine what features are appropriate for specific tasks. For example, in the study of gene expression patterns in Drosophila melanogaster, texture features based on wavelets were particularly effective for determining the developmental stages from in situ hybridization (ISH) images. Such image representation is however not suitable for controlled vocabulary (CV) term annotation because each CV term is often associated with only a part of an image. Here, we developed problem-independent feature extraction methods to generate hierarchical representations for ISH images. Our approach is based on the deep convolutional neural networks (CNNs) that can act on image pixels directly. To make the extracted features generic, the models were trained using a natural image set with millions of labeled examples. These models were transferred to the ISH image domain and used directly as feature extractors to compute image representations. Furthermore, we employed multi-task learning method to fine-tune the pre-trained models with labeled ISH images, and also extracted features from the fine-tuned models. Experimental results showed that feature representations computed by deep models based on transfer and multi-task learning significantly outperformed other methods for annotating gene expression patterns at different stage ranges. We also demonstrated that the intermediate layers of deep models produced the best gene expression pattern representations.
引用
收藏
页码:1475 / 1484
页数:10
相关论文
共 29 条
  • [1] Representation Learning: A Review and New Perspectives
    Bengio, Yoshua
    Courville, Aaron
    Vincent, Pascal
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) : 1798 - 1828
  • [2] Bishop CM, 1995, Neural Networks for Pattern Recognition
  • [3] Collobert Ronan, 2008, INT C MACHINE LEARNI, P160, DOI DOI 10.1145/1390156.1390177
  • [4] Donahue J., 2012, P 31 INT C MACH LEAR, P647
  • [5] Systematic image-driven analysis of the spatial Drosophila embryonic expression landscape
    Frise, Erwin
    Hammonds, Ann S.
    Celniker, Susan E.
    [J]. MOLECULAR SYSTEMS BIOLOGY, 2010, 6
  • [6] Girshick R., 2014, P 27 IEEE C COMP VIS
  • [7] Ji S., 2012, BMC BIOINFORMATICS, V10, P119
  • [8] Automated annotation of Drosophila gene expression patterns using a controlled vocabulary
    Ji, Shuiwang
    Sun, Liang
    Jin, Rong
    Kumar, Sudhir
    Ye, Jieping
    [J]. BIOINFORMATICS, 2008, 24 (17) : 1881 - 1888
  • [9] Ji SW, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P407
  • [10] Krizhevsky Alex., ADV NEURAL INFORM PR, V25, P1106