End-to-end face parsing via interlinked convolutional neural networks

被引:22
|
作者
Yin, Zi [1 ]
Yiu, Valentin [2 ,3 ]
Hu, Xiaolin [2 ]
Tang, Liang [1 ]
机构
[1] Beijing Forestry Univ, Sch Technol, Beijing 100083, Peoples R China
[2] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Inst Artificial Intelligence,THBI,Dept Comp Sci &, Beijing 100084, Peoples R China
[3] Cent Supelec, F-91190 Gif Sur Yvette, France
基金
中国国家自然科学基金;
关键词
STN-iCNN; Face parsing; End-to-end;
D O I
10.1007/s11571-020-09615-4
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Face parsing is an important computer vision task that requires accurate pixel segmentation of facial parts (such as eyes, nose, mouth, etc.), providing a basis for further face analysis, modification, and other applications. Interlinked Convolutional Neural Networks (iCNN) was proved to be an effective two-stage model for face parsing. However, the original iCNN was trained separately in two stages, limiting its performance. To solve this problem, we introduce a simple, end-to-end face parsing framework: STN-aided iCNN(STN-iCNN), which extends the iCNN by adding a Spatial Transformer Network (STN) between the two isolated stages. The STN-iCNN uses the STN to provide a trainable connection to the original two-stage iCNN pipeline, making end-to-end joint training possible. Moreover, as a by-product, STN also provides more precise cropped parts than the original cropper. Due to these two advantages, our approach significantly improves the accuracy of the original model. Our model achieved competitive performance on the Helen Dataset, the standard face parsing dataset. It also achieved superior performance on CelebAMask-HQ dataset, proving its good generalization. Our code has been released at https://github.com/aod321/STN-iCNN.
引用
收藏
页码:169 / 179
页数:11
相关论文
共 50 条
  • [41] End-to-End Image Super-Resolution via Deep and Shallow Convolutional Networks
    Wang, Yifan
    Wang, Lijun
    Wang, Hongyu
    Li, Peihua
    IEEE ACCESS, 2019, 7 : 31959 - 31970
  • [42] End-to-End Emotion Semantic Parsing
    Jiang, Xiaotong
    Wang, Zhongqing
    Zhou, Guodong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 37 - 47
  • [43] End-to-End Exposure Fusion Using Convolutional Neural Network
    Wang, Jinhua
    Wang, Weiqiang
    Xu, Guangmei
    Liu, Hongzhe
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2018, E101D (02): : 560 - 563
  • [44] LEARNING ENVIRONMENTAL SOUNDS WITH END-TO-END CONVOLUTIONAL NEURAL NETWORK
    Tokozume, Yuji
    Harada, Tatsuya
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2721 - 2725
  • [45] End-to-End Hardware Accelerator for Deep Convolutional Neural Network
    Chang, Tian-Sheuan
    2018 INTERNATIONAL SYMPOSIUM ON VLSI DESIGN, AUTOMATION AND TEST (VLSI-DAT), 2018,
  • [46] End-to-end Prediction of Driver Intention using 3D Convolutional Neural Networks
    Gebert, Patrick
    Roitberg, Alina
    Haurilet, Monica
    Stiefelhagen, Rainer
    2019 30TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV19), 2019, : 969 - 974
  • [47] Remote Sensing Airport Detection Based on End-to-End Deep Transferable Convolutional Neural Networks
    Li, Shuai
    Xu, Yuelei
    Zhu, Mingming
    Ma, Shiping
    Tang, Hong
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2019, 16 (10) : 1640 - 1644
  • [48] An End-to-End Steel Strip Surface Defects Recognition System Based on Convolutional Neural Networks
    Yi, Li
    Li, Guangyao
    Jiang, Mingming
    STEEL RESEARCH INTERNATIONAL, 2017, 88 (02) : 176 - 187
  • [49] Virtual Molecular Projections and Convolutional Neural Networks for the End-to-End Modeling of Nanoparticle Activities and Properties
    Russo, Daniel P.
    Yan, Xiliang
    Shende, Sunil
    Huang, Heng
    Yan, Bing
    Zhu, Hao
    ANALYTICAL CHEMISTRY, 2020, 92 (20) : 13971 - 13979
  • [50] Feature map size selection for fMRI classification on end-to-end deep convolutional neural networks
    Suhaimi, Farahana
    Htike, Zaw Zaw
    INTERNATIONAL JOURNAL OF ADVANCED AND APPLIED SCIENCES, 2018, 5 (08): : 95 - 103