Knowledge-enhanced graph convolutional neural networks for text classification

被引:1
作者
Wang T. [1 ]
Zhu X.-F. [1 ]
Tang G. [1 ]
机构
[1] College of Computer Science and Engineering, Chongqing University of Technology, Chongqing
来源
Zhejiang Daxue Xuebao (Gongxue Ban)/Journal of Zhejiang University (Engineering Science) | 2022年 / 56卷 / 02期
关键词
Graph convolutional network; Knowledge embedding; Natural language processing; Neural network; Text classification;
D O I
10.3785/j.issn.1008-973X.2022.02.013
中图分类号
学科分类号
摘要
A new knowledge-enhanced graph convolutional neural network (KEGCN) classification model was proposed aiming at the problem of text classification. In the KEGCN model, firstly a text graph containing word nodes, document nodes, and external entity nodes was constructed on the entire text set. Different similarity calculation methods were used between different types of nodes. After the text graph was constructed, it was input into the two-layer graph convolutional network to learn the representation of the node and classified. The KEGCN model introduced external knowledge to compose the graph, and captured the long-distance discontinuous global semantic information, and was the first work to introduce knowledge information into the graph convolution network for classification tasks. Text classification experiments were conducted on four large-scale real data sets, 20NG, OHSUMED, R52 and R8, and results showed that the classification accuracy of the KEGCN network model was better than that of all baseline models. Results show that integrating knowledge information into the graph convolutional neural network is conducive to learning more accurate text representations and improving the accuracy of text classification. Copyright ©2022 Journal of Zhejiang University (Engineering Science). All rights reserved.
引用
收藏
页码:322 / 328
页数:6
相关论文
共 26 条
[1]  
YAO L, MAO C, LUO Y., Graph convolutional networks for text classification, Proceedings of the AAAI Conference on Artificial Intelligence, 33, 1, pp. 7370-7377, (2019)
[2]  
ZHANG Y, JIN R, ZHOU Z H., Understanding bag-of-words model: a statistical framework, International Journal of Machine Learning and Cybernetics, 1, pp. 43-52, (2010)
[3]  
WANG S I, MANNING C D., Baselines and bigrams: simple, good sentiment and topic classification, Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics, pp. 90-94, (2012)
[4]  
KIM Y., Convolutional neural networks for sentence classification, Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, pp. 1746-1751, (2014)
[5]  
LIU P, QIU X, HUANG X., Recurrent neural network for text classification with multi-task learning, Proceedings of the 25th International Joint Conference on Artificial Intelligence, pp. 2873-2879, (2016)
[6]  
COVER T, HART P., Nearest neighbor pattern classification, IEEE Transactions on Information Theory, 13, 1, pp. 21-27, (1967)
[7]  
UTGOFF P E., ID5: an incremental ID3, Machine Learning Proceedings 1988, pp. 107-120, (1988)
[8]  
LOH W Y., Classification and regression trees, Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 1, 1, pp. 14-23, (2011)
[9]  
QUINLAN J R., C4.5: programs for machine learning, (1993)
[10]  
VATEEKUL P, KUBAT M., Fast induction of multiple decision trees in text categorization from large scale, imbalanced, and multi-label data, 2009 IEEE International Conference on Data Mining Workshops, pp. 320-325, (2009)