Multi-label classification of traditional national costume pattern image semantic understanding

被引:0
作者
Zhao H.-Y. [1 ]
Zhou W. [2 ]
Hou X.-G. [3 ]
Qi G.-L. [4 ]
机构
[1] School of Computer Science, Beijing University of Posts and Telecommunications, Beijing
[2] School of Digital Media and Design Art, Beijing University of Posts and Telecommunications, Beijing
[3] Research Institute of Network Technology, Beijing University of Posts and Telecommunications, Beijing
[4] Century College, Beijing University of Posts and Telecommunications, Beijing
来源
Guangxue Jingmi Gongcheng/Optics and Precision Engineering | 2020年 / 28卷 / 03期
关键词
Fusion learning; Multi-label classification; Semantic understanding; Traditional national costumes;
D O I
10.3788/OPE.20202803.0695
中图分类号
学科分类号
摘要
Since current image multi-label classification methods only focus on the category information of image ontology (ontology) and ignore the deep semantic information of the image (implicit), this study proposed an image multi-label classification model of "ontology-implicit" fusion learning. The model first used the middle and higher layers of CNN to learn the image ontology information and implicit information, respectively, and then it used the dependency relationship between the ontology information and implicit information to design the fusion learning model. Meanwhile, the different characteristics of the middle layer and different structures of the model were studied in-depth, to realize the classification of implicit information contained in multiple image categories. Experiments conducted on the traditional national costume pattern image datasets show that the mAP of image ontology multi-label classification and implicit multi-label classification are 0.88 and 0.82, respectively. Comparative experiments conducted on the Scene dataset show that the model is superior to other methods in Hamming loss, one error, and average precision indices, with values of 0.103, 0.091, and 0.083, respectively. Therefore, the experimental results prove the effectiveness and superiority of this method. © 2020, Science Press. All right reserved.
引用
收藏
页码:695 / 703
页数:8
相关论文
共 20 条
[1]  
Zhang K.H., Tan Z.H., Li B., Automated image segmentation based on pulse coupled neural network with partide swarm optimization and comprehensive evaluation, Opt. Precision Eng, 26, 4, pp. 962-970, (2018)
[2]  
Liu Z., Huang J.T., Feng X., Action recognition model construction based on multi-scale deep convolution neural network, Opt. Precision Eng., 25, 3, pp. 799-805, (2017)
[3]  
Li Y., Liu X.Y., Zhang H.Q., Et al., Optical remote sensing image retrieval based on convolutional neural networks, Opt. Precision Eng., 26, 1, pp. 200-207, (2018)
[4]  
Wei Y., Xia W., Lin M., Et al., HCP: A flexible CNN framework for multi-label image classification, IEEE Transactions on Pattern Analysis and Machine Intelligence, 38, 9, pp. 1901-1907, (2015)
[5]  
Wang Z., Chen T., Li G., Et al., Multi-label image recognition by recurrently discovering attentional regions, Proceedings of the IEEE International Conference on Computer Vision (CVPR), pp. 464-472, (2017)
[6]  
Yu W.J., Chen Z.D., Luo X., Et al., DELTA: A deep dual-stream network for multi-label image classification, Pattern Recognition, 91, pp. 322-331, (2019)
[7]  
Yan Z., Liu W.W., Wen S.P., Et al., Multi-label image classification by feature attention network, IEEE Access, 7, pp. 98005-98013, (2019)
[8]  
Song P., Jing L.P., Et al., Exploiting label relationships in multi-label classification with neural networks, Journal of Computer Research and Development, 55, 8, pp. 1751-1759, (2018)
[9]  
Jin J., Nakayama H., Annotation order matters: Recurrent image annotator for arbitrary length image tagging, 201623rd International Conference on Pattern Recognition (ICPR), pp. 2452-2457, (2016)
[10]  
Wang J., Yang Y., Mao J., Et al., Cnn-rnn: A unified framework for multi-label image classification, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2285-2294, (2016)