Dual Perspective of Label-Specific Feature Learning for Multi-Label Classification

被引:0
作者
Hang, Jun-Yi [1 ]
Zhang, Min-Ling [1 ]
机构
[1] Southeast University, Nanjing
基金
中国国家自然科学基金;
关键词
label-specific features; Machine learning; missing labels; multi-label classification; partial labels;
D O I
10.1145/3705006
中图分类号
学科分类号
摘要
Label-specific features work as an effective supervised feature manipulation strategy to account for distinct discriminative properties of each class label in multi-label classification. Existing approaches implement this strategy in its primal form, i.e., finding the most pertinent features specific to each class label and directly inducing classifiers on these features. Instead of such a straightforward implementation, a dual perspective for label-specific feature learning is investigated in this article. As a dual problem of existing primal one, we consider label-specific discriminative properties by identifying non-informative features for each class label and making the discrimination process immutable to variations of identified features. Accordingly, a perturbation-based approach Dela is presented, which endows classifiers with immutability on simultaneously identified non-informative features by solving a probabilistically relaxed expected risk minimization problem. Furthermore, we touch the realistic issue of label-specific feature learning in a weakly supervised scenario via extending Dela to accommodate to multi-label data with missing labels. Comprehensive experiments show that our approach outperforms the state-of-the-art counterparts. © 2024 Copyright held by the owner/author(s).
引用
收藏
相关论文
共 98 条
  • [1] Achille A., Soatto S., Information dropout: Learning optimal representations through noisy computation, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40, 12, pp. 2897-2905, (2018)
  • [2] Ahuja K., Caballero E., Zhang D., Gagnon-Audet J.-C., Bengio Y., Mitliagkas I., Rish I., Invariance principle meets information bottleneck for out-of-distribution generalization, Proceedings of the 35th International Conference on Neural Information Processing Systems, pp. 3438-3450, (2021)
  • [3] Ahuja K., Shanmugam K., Varshney K.R., Dhurandhar A., Invariant risk minimization games, Proceedings of the 37th International Conference on Machine Learning, pp. 145-155, (2020)
  • [4] Alemi A.A., Fischer I., Dillon J.V., Murphy K., Deep variational information bottleneck, Proceedings of the 5th International Conference on Learning Representations, (2017)
  • [5] Arazo E., Ortego D., Albert P., O'Connor N.E., McGuinness K., Pseudo-labeling and confirmation bias in deep semi-supervised learning, Proceedings of the International Joint Conference on Neural Networks, pp. 1-8, (2020)
  • [6] Arpit D., Jastrzebski S., Ballas N., Krueger D., Bengio E., Kanwal M.S., Maharaj T., Fischer A., Courville A.C., Bengio Y., Et al., A closer look at memorization in deep networks, Proceedings of the 34th International Conference on Machine Learning, pp. 233-242, (2017)
  • [7] Bai J., Kong S., Gomes C.P., Disentangled variational autoencoder based multi-label classification with covariance-aware multivariate probit model, Proceedings of the 29th International Joint Conference on Artificial Intelligence, pp. 4313-4321, (2020)
  • [8] Baruch E.B., Ridnik T., Friedman I., Ben-Cohen A., Zamir N., Noy A., Zelnik-Manor L., Multi-label classification with partial annotations using class-aware selective loss, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4754-4762, (2022)
  • [9] Blum A., Haghtalab N., Procaccia A.D., Variational dropout and the local reparameterization trick, Proceedings of the 28th International Conference on Neural Information Processing Systems, pp. 2575-2583, (2015)
  • [10] Boutell M.R., Luo J.-B., Shen X.-P., Brown C.M., Learning multi-label scene classification, Pattern Recognition, 37, 9, pp. 1757-1771, (2004)