Prototype Memory for Large-Scale Face Representation Learning

被引:3
作者
Smirnov, Evgeny [1 ]
Garaev, Nikita [1 ]
Galyuk, Vasiliy [1 ]
Lukyanets, Evgeny [2 ]
机构
[1] Speech Technol Ctr, St Petersburg 194044, Russia
[2] ITMO Univ, Dept Informat Technol & Programming, St Petersburg 197101, Russia
关键词
Prototypes; Face recognition; Training; Random access memory; Graphics processing units; Representation learning; Task analysis; Deep neural networks; face recognition; representation learning; softmax acceleration; RECOGNITION;
D O I
10.1109/ACCESS.2022.3146059
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Face representation learning using datasets with a massive number of identities requires appropriate training methods. Softmax-based approach, currently the state-of-the-art in face recognition, in its usual "full softmax" form is not suitable for datasets with millions of persons. Several methods, based on the "sampled softmax" approach, were proposed to remove this limitation. These methods, however, have a set of disadvantages. One of them is a problem of "prototype obsolescence": classifier weights (prototypes) of the rarely sampled classes receive too scarce gradients and become outdated and detached from the current encoder state, resulting in incorrect training signals. This problem is especially serious in ultra-large-scale datasets. In this paper, we propose a novel face representation learning model called Prototype Memory, which alleviates this problem and allows training on a dataset of any size. Prototype Memory consists of the limited-size memory module for storing recent class prototypes and employs a set of algorithms to update it in appropriate way. New class prototypes are generated on the fly using exemplar embeddings in the current mini-batch. These prototypes are enqueued to the memory and used in a role of classifier weights for softmax classification-based training. To prevent obsolescence and keep the memory in close connection with the encoder, prototypes are regularly refreshed, and oldest ones are dequeued and disposed of. Prototype Memory is computationally efficient and independent of dataset size. It can be used with various loss functions, hard example mining algorithms and encoder architectures. We prove the effectiveness of the proposed model by extensive experiments on popular face recognition benchmarks.
引用
收藏
页码:12031 / 12046
页数:16
相关论文
共 141 条
[1]  
Ahmed Khan S., 2021, ARXIV
[2]   On the frontiers of pose invariant face recognition: a review [J].
Ahmed, Sheikh Bilal ;
Ali, Syed Farooq ;
Ahmad, Jameel ;
Adnan, Muhammad ;
Fraz, Muhammad Moazam .
ARTIFICIAL INTELLIGENCE REVIEW, 2020, 53 (04) :2571-2634
[3]  
Ali A., 2020, P EUR C COMP VIS CHA, P133
[4]  
An X., ARXIV201005222
[5]  
Bai Y., 2017, ARXIV170703073
[6]  
Bamler R., 2020, P ICLR
[7]  
Bansal A, 2017, 2017 IEEE INTERNATIONAL JOINT CONFERENCE ON BIOMETRICS (IJCB), P456
[8]  
Blanc G, 2018, PR MACH LEARN RES, V80
[9]  
Boutros F., 2021, UNMASKING FACE EMBED
[10]   Albumentations: Fast and Flexible Image Augmentations [J].
Buslaev, Alexander ;
Iglovikov, Vladimir I. ;
Khvedchenya, Eugene ;
Parinov, Alex ;
Druzhinin, Mikhail ;
Kalinin, Alexandr A. .
INFORMATION, 2020, 11 (02)