Improving the Representation of CNN Based Features by Autoencoder for a Task of Construction Material Image Classification

被引:16
作者
Bunrit, S. [1 ]
Kerdprasop, N. [1 ]
Kerdprasop, K. [1 ]
机构
[1] Suranaree Univ Technol, Suranari, Thailand
关键词
Convolution Neural Network (CNN); transfer learning; Autoencoder; construction material; image classification; RECOGNITION;
D O I
10.12720/jait.11.4.192-199
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning based model named Convolution Neural Network (CNN) has been extensively employed by diversified applications concerned images or videos data. Because training a specific CNN model for an application task consumes enormous machine resources and need many of the training data, consequently pre-trained models of CNN have been broadly used as the transfer-learning scenario. By the scenario, features had been learned from a pre-trained model by one source task can be proficiently sent further to another specific task in a concept of knowledge transferring. As a result, a task specific can be directly employed such pre-trained features or further train more by setting the pre-trained features as a starting point. Thereby, it takes not much time and can improve the performance from many referenced works. In this work, with a task specific on construction material images classification, we investigate on the transfer learning of GoogleNet and ResNet101 that pre-trained on ImageNet dataset (source task). By applying both of the transfer-learning schemes, they reveal quite satisfied results. The best for GoogleNet, it gets 95.50 percent of the classification accuracy by fine-tuning scheme. Where, for ResNet101, the best is of 95.00 percent by using fixed feature extractor scheme. Nevertheless, after the learning based representation methods are further employed on top of the transferred features, they expose more appeal results. By Autoencoder based representation method reveals the performance can improve more than PCA (Principal Component Analysis) in all cases. Especially, when the fixed feature extractor of ResNet101 is used as the input to Autoencoder, the classified result can be improved up to 97.83%. It can be inferred, just applying Autoencoder on top of the pre-trained transferred features, the performance can be improved by we have no need to fine-tune the complex pre-trained model.
引用
收藏
页码:192 / 199
页数:8
相关论文
共 18 条
  • [1] Construction site image retrieval based on material cluster recognition
    Brilakis, Ioannis K.
    Soibelman, Lucio
    Shinagawa, Yoshihisa
    [J]. ADVANCED ENGINEERING INFORMATICS, 2006, 20 (04) : 443 - 452
  • [2] Bridging BIM and building: From a literature review to an integrated conceptual framework
    Chen, Ke
    Lu, Weisheng
    Peng, Yi
    Rowlinson, Steve
    Huang, George Q.
    [J]. INTERNATIONAL JOURNAL OF PROJECT MANAGEMENT, 2015, 33 (06) : 1405 - 1416
  • [3] Christian S., 2015, P IEEE C COMP VIS PA, DOI [DOI 10.1109/CVPR.2015.7298594, 10.1109/CVPR.2015.7298594]
  • [4] cs231n.github.io, TRANSFER LEARNING
  • [5] Dahan T. M., 2017, 6220 INSE CONC U
  • [6] Geometry-Informed Material Recognition
    Degol, Joseph
    Golparvar-Fard, Mani
    Hoiem, Derek
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1554 - 1562
  • [7] Vision-based material recognition for automated monitoring of construction progress and generating building information modeling from unordered site image collections
    Dimitrov, Andrey
    Golparvar-Fard, Mani
    [J]. ADVANCED ENGINEERING INFORMATICS, 2014, 28 (01) : 37 - 49
  • [8] Elhabian S, 2009, TUTORIAL DATA REDUCT
  • [9] Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
  • [10] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778