Novel Efficient RNN and LSTM-Like Architectures: Recurrent and Gated Broad Learning Systems and Their Applications for Text Classification

被引:142
作者
Du, Jie [1 ]
Vong, Chi-Man [2 ]
Chen, C. L. Philip [3 ,4 ]
机构
[1] Shenzhen Univ, Sch Biomed Engn,Hlth Sci Ctr, Natl Reg Key Technol Engn Lab Med Ultrasound, Guangdong Key Lab Biomed Measurements & Ultrasoun, Shenzhen 518060, Peoples R China
[2] Univ Macau, Dept Comp & Informat Sci, Macau, Peoples R China
[3] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510641, Peoples R China
[4] Univ Macau, Fac Sci & Technol, Macau, Peoples R China
基金
中国国家自然科学基金;
关键词
Logic gates; Learning systems; Training; Computer architecture; Recurrent neural networks; Task analysis; Broad learning system (BLS); sequence information; simultaneous learning; text classification; word importance; NEURAL-NETWORKS; MACHINE;
D O I
10.1109/TCYB.2020.2969705
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
High accuracy of text classification can be achieved through simultaneous learning of multiple information, such as sequence information and word importance. In this article, a kind of flat neural networks called the broad learning system (BLS) is employed to derive two novel learning methods for text classification, including recurrent BLS (R-BLS) and long short-term memory (LSTM)-like architecture: gated BLS (G-BLS). The proposed two methods possess three advantages: 1) higher accuracy due to the simultaneous learning of multiple information, even compared to deep LSTM that extracts deeper but single information only; 2) significantly faster training time due to the noniterative learning in BLS, compared to LSTM; and 3) easy integration with other discriminant information for further improvement. The proposed methods have been evaluated over 13 real-world datasets from various types of text classification. From the experimental results, the proposed methods achieve higher accuracies than LSTM while taking significantly less training time on most evaluated datasets, especially when the LSTM is in deep architecture. Compared to R-BLS, G-BLS has an extra forget gate to control the flow of information (similar to LSTM) to further improve the accuracy on text classification so that G-BLS is more effective while R-BLS is more efficient.
引用
收藏
页码:1586 / 1597
页数:12
相关论文
共 40 条
[31]   Bidirectional recurrent neural networks [J].
Schuster, M ;
Paliwal, KK .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1997, 45 (11) :2673-2681
[32]   Principal Component 2-D Long Short-Term Memory for Font Recognition on Single Chinese Characters [J].
Tao, Dapeng ;
Lin, Xu ;
Jin, Lianwen ;
Li, Xuelong .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (03) :756-765
[33]  
Tulyakov S, 2008, STUD COMPUT INTELL, V90, P361
[34]  
Venugopalan S., 2015, 2015 ANN C N AM CHAP, P1494
[35]   Toward a New Task Assignment and Path Evolution (TAPE) for Missile Defense System (MDS) Using Intelligent Adaptive SOM with Recurrent Neural Networks (RNNs) [J].
Wang, Chi-Hsu ;
Chen, Chun-Yao ;
Hung, Kun-Neng .
IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (06) :1134-1145
[36]   BACKPROPAGATION THROUGH TIME - WHAT IT DOES AND HOW TO DO IT [J].
WERBOS, PJ .
PROCEEDINGS OF THE IEEE, 1990, 78 (10) :1550-1560
[37]   Neural networks for classification: A survey [J].
Zhang, GQP .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2000, 30 (04) :451-462
[38]  
Zhang X, 2015, ADV NEUR IN, V28
[39]   Drawing and Recognizing Chinese Characters with Recurrent Neural Network [J].
Zhang, Xu-Yao ;
Yin, Fei ;
Zhang, Yan-Ming ;
Liu, Cheng-Lin ;
Bengio, Yoshua .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :849-862
[40]   Triplex Transfer Learning: Exploiting Both Shared and Distinct Concepts for Text Classification [J].
Zhuang, Fuzhen ;
Luo, Ping ;
Du, Changying ;
He, Qing ;
Shi, Zhongzhi ;
Xiong, Hui .
IEEE TRANSACTIONS ON CYBERNETICS, 2014, 44 (07) :1191-1203