Class-Conditional Sharpness-Aware Minimization for Deep Long-Tailed Recognition

被引:17
|
作者
Zhou, Zhipeng [1 ]
Li, Lanqing [2 ,4 ]
Zhao, Peilin [3 ]
Heng, Pheng-Ann [2 ]
Gong, Wei [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[3] Tencent AI Lab, Shenzhen, Peoples R China
[4] Zhejiang Lab, Hangzhou, Peoples R China
关键词
D O I
10.1109/CVPR52729.2023.00341
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It's widely acknowledged that deep learning models with flatter minima in its loss landscape tend to generalize better. However, such property is under-explored in deep long-tailed recognition (DLTR), a practical problem where the model is required to generalize equally well across all classes when trained on highly imbalanced label distribution. In this paper, through empirical observations, we argue that sharp minima are in fact prevalent in deep longtailed models, whereas naive integration of existing flattening operations into long-tailed learning algorithms brings little improvement. Instead, we propose an effective two-stage sharpness-aware optimization approach based on the decoupling paradigm in DLTR. In the first stage, both the feature extractor and classifier are trained under parameter perturbations at a class-conditioned scale, which is theoretically motivated by the characteristic radius of flat minima under the PAC-Bayesian framework. In the second stage, we generate adversarial features with class-balanced sampling to further robustify the classifier with the backbone frozen. Extensive experiments on multiple longtailed visual recognition benchmarks show that, our proposed Class-Conditional Sharpness-Aware Minimization (CC-SAM), achieves competitive performance compared to the state-of-the-arts. Code is available at https://github.com/zzpustc/CC-SAM.
引用
收藏
页码:3499 / 3509
页数:11
相关论文
共 50 条
  • [1] Sharpness-Aware Model-Agnostic Long-Tailed Domain Generalization
    Su, Houcheng
    Luo, Weihao
    Liu, Daixian
    Wang, Mengzhu
    Tang, Jing
    Chen, Junyang
    Wang, Cong
    Chen, Zhenghan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 15091 - 15099
  • [2] ImbSAM: A Closer Look at Sharpness-Aware Minimization in Class-Imbalanced Recognition
    Zhou, Yixuan
    Qu, Yi
    Xu, Xing
    Shen, Hengtao
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11311 - 11321
  • [3] Random Sharpness-Aware Minimization
    Liu, Yong
    Mai, Siqi
    Cheng, Minhao
    Chen, Xiangning
    Hsieh, Cho-Jui
    You, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [4] Friendly Sharpness-Aware Minimization
    Li, Tao
    Zhou, Pan
    He, Zhengbao
    Cheng, Xinwen
    Huang, Xiaolin
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 5631 - 5640
  • [5] Convergence of Sharpness-Aware Minimization with Momentum
    Pham Duy Khanh
    Luong, Hoang-Chau
    Mordukhovich, Boris S.
    Dat Ba Tran
    Truc Vo
    INFORMATION TECHNOLOGIES AND THEIR APPLICATIONS, PT II, ITTA 2024, 2025, 2226 : 123 - 132
  • [6] Towards Understanding Sharpness-Aware Minimization
    Andriushchenko, Maksym
    Flammarion, Nicolas
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 639 - 668
  • [7] Sharpness-Aware Minimization and the Edge of Stability
    Long, Philip M.
    Bartlett, Peter L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 20
  • [8] Noise-resistant sharpness-aware minimization in deep learning
    Su, Dan
    Jin, Long
    Wang, Jun
    NEURAL NETWORKS, 2025, 181
  • [9] Towards Efficient and Scalable Sharpness-Aware Minimization
    Liu, Yong
    Mai, Siqi
    Chen, Xiangning
    Hsieh, Cho-Jui
    You, Yang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12350 - 12360
  • [10] The Crucial Role of Normalization in Sharpness-Aware Minimization
    Dai, Yan
    Ahn, Kwangjun
    Sra, Suvrit
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,