TransBLS: transformer combined with broad learning system for facial beauty prediction

被引:4
作者
Gan, Junying [1 ]
Xie, Xiaoshan [1 ]
He, Guohui [1 ]
Luo, Heng [1 ]
机构
[1] Wuyi Univ, Dept Intelligent Mfg, Jiangmen 529020, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Adaptive transformer; Broad learning system; Facial beauty prediction; Global self-attention; Local self-attention;
D O I
10.1007/s10489-023-04931-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial beauty prediction (FBP) is a frontier topic in the fields of machine learning and computer vision, focusing on how to enable computers to judge facial beauty like humans. The existing FBP methods are mainly based on deep neural networks (DNNs). However, DNNs lack global characteristics and only build local dependencies, so FBP still suffers from insufficient supervision information, low accuracy and overfitting. A transformer is a self-attention-based architecture that possesses better global characteristics than DNNs and can build long-term dependencies. Transformers have been widely used to solve some computer vision problems in recent years and have produced better results. In this paper, we propose an adaptive transformer with global and local multihead self-attention for FBP, called GLAFormer. However, GLAFormer does not converge and is prone to overfitting when the training samples are insufficient. The broad learning system (BLS) can accelerate the model convergence process and reduce overfitting. Therefore, we further combine GLAFormer with the BLS to form TransBLS, in which a GLAFormer block is designed as a feature extractor, the features extracted by it are transferred to the BLS for further refining and fitting, and the results are output. Experimental results indicate that TransBLS achieves state-of-the-art FBP performance on several datasets with different scales, better solving the low accuracy and overfitting problems encountered in FBP. It can also be widely applied in pattern recognition and object detection tasks.
引用
收藏
页码:26110 / 26125
页数:16
相关论文
共 50 条
[1]   A Unified Model for Face Matching and Presentation Attack Detection using an Ensemble of Vision Transformer Features [J].
Al-Refai, Rouqaiah ;
Nandakumar, Karthik .
2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WORKSHOPS (WACVW), 2023, :662-671
[2]   Deep learning based face beauty prediction via dynamic robust losses and ensemble regression [J].
Bougourzi, F. ;
Dornaika, F. ;
Taleb-Ahmed, A. .
KNOWLEDGE-BASED SYSTEMS, 2022, 242
[3]   CNN based facial aesthetics analysis through dynamic robust losses and ensemble regression [J].
Bougourzi, Fares ;
Dornaika, Fadi ;
Barrena, Nagore ;
Distante, Cosimo ;
Taleb-Ahmed, Abdelmalik .
APPLIED INTELLIGENCE, 2023, 53 (09) :10825-10842
[4]   Deep Learning for Facial Beauty Prediction [J].
Cao, Kerang ;
Choi, Kwang-nam ;
Jung, Hoekyung ;
Duan, Lini .
INFORMATION, 2020, 11 (08)
[5]   Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture [J].
Chen, C. L. Philip ;
Liu, Zhulin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (01) :10-24
[6]   Data-Driven Facial Beauty Analysis: Prediction, Retrieval and Manipulation [J].
Chen, Fangmei ;
Xiao, Xihua ;
Zhang, David .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2018, 9 (02) :205-216
[7]  
Chen SF, 2022, ADV NEUR IN
[8]   A New Automatic Hyperparameter Recommendation Approach Under Low-Rank Tensor Completion e Framework [J].
Deng, Liping ;
Xiao, Mingqing .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) :4038-4050
[9]  
Dosovitskiy A., 2021, P INT C LEARN REPR, DOI DOI 10.48550/ARXIV.2010.11929
[10]  
Fan Yang, 2018, 2018 IEEE 4th International Conference on Computer and Communications (ICCC). Proceedings, P2105, DOI 10.1109/CompComm.2018.8780984