Facial landmark points detection using knowledge distillation-based neural networks

被引:14
作者
Fard, Ali Pourramezan [1 ]
Mahoor, Mohammad H. [1 ]
机构
[1] Univ Denver, Dept Elect & Comp Engn, 2155 E Wesley Ave, Denver, CO 80208 USA
关键词
Deep learning; Face alignment; Facial landmark points detection; Knowledge distillation; FACE ALIGNMENT;
D O I
10.1016/j.cviu.2021.103316
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial landmark detection is a vital step for numerous facial image analysis applications. Although some deep learning-based methods have achieved good performances in this task, they are often not suitable for running on mobile devices. Such methods rely on networks with many parameters, which makes the training and inference time-consuming. Training lightweight neural networks such as MobileNets are often challenging, and the models might have low accuracy. Inspired by knowledge distillation (KD), this paper presents a novel loss function to train a lightweight Student network (e.g., MobileNetV2) for facial landmark detection. We use two Teacher networks, a Tolerant-Teacher and a Tough-Teacher in conjunction with the Student network. The Tolerant-Teacher is trained using Soft-landmarks created by active shape models, while the Tough-Teacher is trained using the ground truth (aka Hard-landmarks) landmark points. To utilize the facial landmark points predicted by the Teacher networks, we define an Assistive Loss (ALoss) for each Teacher network. Moreover, we define a loss function called KD-Loss that utilizes the facial landmark points predicted by the two pre-trained Teacher networks (EfficientNet-b3) to guide the lightweight Student network towards predicting the Hard landmarks. Our experimental results on three challenging facial datasets show that the proposed architecture will result in a better-trained Student network that can extract facial landmark points with high accuracy.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Bi-Level Multi-column Convolutional Neural Networks for Facial Landmark Point Detection
    Xu, Yanyu
    Gao, Shenghua
    [J]. COMPUTER VISION - ECCV 2016 WORKSHOPS, PT II, 2016, 9914 : 536 - 551
  • [42] Knowledge distillation-based information sharing for online process monitoring in decentralized manufacturing system
    Shi, Zhangyue
    Li, Yuxuan
    Liu, Chenang
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2025, 36 (03) : 2177 - 2192
  • [43] Knowledge Distillation-Based Representation Learning for Short-Utterance Spoken Language Identification
    Shen, Peng
    Lu, Xugang
    Li, Sheng
    Kawai, Hisashi
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 (28) : 2674 - 2683
  • [44] Melanoma Breslow Thickness Classification Using Ensemble-Based Knowledge Distillation With Semi-Supervised Convolutional Neural Networks
    Dominguez-Morales, Juan P.
    Hernandez-Rodriguez, Juan-Carlos
    Duran-Lopez, Lourdes
    Conejo-Mir, Julian
    Pereyra-Rodriguez, Jose-Juan
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (01) : 443 - 455
  • [45] Knowledge distillation with ensembles of convolutional neural networks for medical image segmentation
    Noothout, Julia M. H.
    Lessmann, Nikolas
    van Eede, Matthijs C.
    van Harten, Louis D.
    Sogancioglu, Ecem
    Heslinga, Friso G.
    Veta, Mitko
    van Ginneken, Bram
    Isgum, Ivana
    [J]. JOURNAL OF MEDICAL IMAGING, 2022, 9 (05)
  • [46] Lightweight Knowledge Distillation-Based Transfer Learning Framework for Rolling Bearing Fault Diagnosis
    Lu, Ruijia
    Liu, Shuzhi
    Gong, Zisu
    Xu, Chengcheng
    Ma, Zonghe
    Zhong, Yiqi
    Li, Baojian
    [J]. SENSORS, 2024, 24 (06)
  • [47] Stacked attention hourglass network based robust facial landmark detection
    Huang, Ying
    Huang, He
    [J]. NEURAL NETWORKS, 2023, 157 : 323 - 335
  • [48] BAG-OF-FEATURES-BASED KNOWLEDGE DISTILLATION FOR LIGHTWEIGHT CONVOLUTIONAL NEURAL NETWORKS
    Chariton, Alexandros
    Passalis, Nikolaos
    Tefas, Anastasios
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1541 - 1545
  • [49] Effective training of convolutional neural networks for age estimation based on knowledge distillation
    Antonio Greco
    Alessia Saggese
    Mario Vento
    Vincenzo Vigilante
    [J]. Neural Computing and Applications, 2022, 34 : 21449 - 21464
  • [50] Effective training of convolutional neural networks for age estimation based on knowledge distillation
    Greco, Antonio
    Saggese, Alessia
    Vento, Mario
    Vigilante, Vincenzo
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (24) : 21449 - 21464