Learning Prototype Classifiers for Long-Tailed Recognition

被引:0
作者
Sharma, Saurabh [1 ]
Xian, Yongqin [2 ]
Yu, Ning [3 ]
Singh, Ambuj [1 ]
机构
[1] Univ Calif Santa Barbara, Santa Barbara, CA 93106 USA
[2] Google, Zurich, Switzerland
[3] Salesforce Res, New York, NY 10001 USA
来源
PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of long-tailed recognition (LTR) has received attention in recent years due to the fundamental power-law distribution of objects in the real-world. Most recent works in LTR use softmax classifiers that are biased in that they correlate classifier norm with the amount of training data for a given class. In this work, we show that learning prototype classifiers addresses the biased softmax problem in LTR. Prototype classifiers can deliver promising results simply using Nearest-Class-Mean (NCM), a special case where prototypes are empirical centroids. We go one step further and propose to jointly learn prototypes by using distances to prototypes in representation space as the logit scores for classification. Further, we theoretically analyze the properties of Euclidean distance based prototype classifiers that lead to stable gradient-based optimization which is robust to outliers. To enable independent distance scales along each channel, we enhance Proto-type classifiers by learning channel-dependent temperature parameters. Our analysis shows that prototypes learned by Prototype classifiers are better separated than empirical centroids. Results on four LTR benchmarks show that Prototype classifier outperforms or is comparable to state-of-the-art methods. Our code is made available at https://github.com/saurabhsharma1993/prototype-classifier-ltr.
引用
收藏
页码:1360 / 1368
页数:9
相关论文
共 39 条
[11]  
DeVries Terrance, 2017, arXiv
[12]  
Drummond Chris, 2003, WORKSH LEARN INB DAT
[13]  
Estabrooks Andrew, 2004, COMPUTATIONAL INTELL
[14]   Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training [J].
Fang, Cong ;
He, Hangfeng ;
Long, Qi ;
Su, Weijie J. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (43)
[15]  
Guerriero S., 2018, INT C LEARN REPR WOR, P1
[16]   Learning from Imbalanced Data [J].
He, Haibo ;
Garcia, Edwardo A. .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2009, 21 (09) :1263-1284
[17]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[18]  
Kang Bingyi, 2019, 8 INT C LEARN REPR I
[19]  
Lin T. Y., 2017, P IEEE C COMPUTER VI, P936, DOI [10.1109/CVPR.2017.106, https://doi.org/10.48550/arXiv.1612.03144, DOI 10.1109/CVPR.2017.106]
[20]   Large-Scale Long-Tailed Recognition in an Open World [J].
Liu, Ziwei ;
Miao, Zhongqi ;
Zhan, Xiaohang ;
Wang, Jiayun ;
Gong, Boqing ;
Yu, Stella X. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2532-2541