Partial Label Learning Based on Fully Connected Deep Neural Network

被引:0
作者
Li H. [1 ]
Wu L. [1 ]
He J. [1 ]
Zheng R. [1 ]
Zhou Y. [1 ]
Qiao S. [2 ]
机构
[1] College of Information and Communication Engineering, Dalian Minzu University, Dalian
[2] School of Management, Dalian Polytechnic University, Dalian
来源
International Journal of Circuits, Systems and Signal Processing | 2022年 / 16卷
基金
中国国家自然科学基金;
关键词
Deep neural network; Partial label learning; Weakly annotated data;
D O I
10.46300/9106.2022.16.35
中图分类号
学科分类号
摘要
The ambiguity of training samples in the partial label learning framework makes it difficult for us to develop learning algorithms and most of the existing algorithms are proposed based on the traditional shallow machine learning models, such as decision tree, support vector machine, and Gaussian process model. Deep neural networks have demonstrated excellent performance in many application fields, but currently it is rarely used for partial label learning frame-work. This study proposes a new partial label learning algorithm based on a fully connected deep neural network, in which the relationship between the candidate labels and the ground-truth label of each training sample is established by defining three new loss functions, and a regularization term is added to prevent overfitting. The experimental results on the controlled U-CI datasets and real-world partial label datasets reveal that the proposed algorithm can achieve higher classification accuracy than the state-of-the-art partial label learning algorithms. © 2022, North Atlantic University Union NAUN. All rights reserved.
引用
收藏
页码:287 / 297
页数:10
相关论文
共 54 条
  • [41] Zhou Y., Gu H., Geometric mean metric learning for partial label data, Neurocomputing, 275, pp. 394-402, (2018)
  • [42] Xu S., Yang M., Zhou Y., Zheng R., Liu W., He J., Partial label metric learning by collapsing classes, International Journal of Machine Learning and Cybernetics, 11, pp. 2453-2460, (2020)
  • [43] Zhou Y., He J., Gu H., Partial label learning via gaussian processes, IEEE Transactions on Cy-bernetics, 47, 12, pp. 4443-4450, (2017)
  • [44] Xing-si L., An aggregate function method for non-linear programming, Science in China Series A-Mathematics, Physics, Astronomy & Technological Science, 34, 12, pp. 1467-1473, (1991)
  • [45] Shi G., Zhang J., Li H., Wang C., Enhance the performance of deep neural networks via l2 regular-ization on the input of activations, Neural Processing Letters, pp. 1-19, (2018)
  • [46] Zheng Q., Fang J., Hu Z., Zhang H., Aero-engine on-board model based on batch normalize deep neural network, IEEE Access, 7, pp. 54855-54862, (2019)
  • [47] Zhang M.-L., Yu F., Tang C.-Z., Disambiguation-free partial label learning, IEEE Transactions on Knowledge and Data Engineering, 29, 10, pp. 2155-2167, (2017)
  • [48] Liu L., Dietterich T. G., A conditional multi-nomial mixture model for superset label learning, Advances in neural information processing systems, pp. 548-556, (2012)
  • [49] Briggs F., Fern X. Z., Raich R., Rank-loss support instance machines for miml instance annotation, Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 534-542, (2012)
  • [50] Han H., Otto C., Liu X., Jain A. K., De-mographic estimation from face images: Human vs. machine performance, IEEE transactions on pattern analysis and machine intelligence, 37, 6, pp. 1148-1161, (2014)