A Unified Understanding of Deep NLP Models for Text Classification

被引:25
作者
Li, Zhen [1 ]
Wang, Xiting [2 ]
Yang, Weikai [1 ]
Wu, Jing [3 ]
Zhang, Zhengyan [4 ]
Liu, Zhiyuan [4 ]
Sun, Maosong [4 ]
Zhang, Hui [1 ]
Liu, Shixia [1 ]
机构
[1] Tsinghua Univ, Sch Software, BNRist, Beijing 100084, Peoples R China
[2] Microsoft Res Asia, Beijing 100080, Peoples R China
[3] Cardiff Univ, Cardiff CF10 3AT, Wales
[4] Tsinghua Univ, Dept Comp Sci & Technol, BNRist, Beijing 100084, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Computational modeling; Analytical models; Visualization; Internet; Data models; Computer architecture; Adaptation models; Explainable AI; visual debugging; visual analytics; deep NLP model; information-based interpretation;
D O I
10.1109/TVCG.2022.3184186
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The rapid development of deep natural language processing (NLP) models for text classification has led to an urgent need for a unified understanding of these models proposed individually. Existing methods cannot meet the need for understanding different models in one framework due to the lack of a unified measure for explaining both low-level (e.g., words) and high-level (e.g., phrases) features. We have developed a visual analysis tool, DeepNLPVis, to enable a unified understanding of NLP models for text classification. The key idea is a mutual information-based measure, which provides quantitative explanations on how each layer of a model maintains the information of input words in a sample. We model the intra- and inter-word information at each layer measuring the importance of a word to the final prediction as well as the relationships between words, such as the formation of phrases. A multi-level visualization, which consists of a corpus-level, a sample-level, and a word-level visualization, supports the analysis from the overall training set to individual samples. Two case studies on classification tasks and comparison between models demonstrate that DeepNLPVis can help users effectively identify potential problems caused by samples and model architectures and then make informed improvements.
引用
收藏
页码:4980 / 4994
页数:15
相关论文
共 49 条
[1]  
[Anonymous], 2016, P 2016 C N AM CHAPTE
[2]  
Baehrens D, 2010, J MACH LEARN RES, V11, P1803
[3]  
Brunner Gino, 2020, 8 INT C LEARNING REP
[4]   RNNbow: Visualizing Learning Via Backpropagation Gradients in RNNs [J].
Cashman, Dylan ;
Patterson, Genevieve ;
Mosca, Abigail ;
Watts, Nathan ;
Robinson, Shannon ;
Chang, Remco .
IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2018, 38 (06) :39-50
[5]   Token-wise sentiment decomposition for ConvNet: Visualizing a sentiment classifier [J].
Chawla, Piyush ;
Hazarika, Subhashis ;
Shen, Han-Wei .
VISUAL INFORMATICS, 2020, 4 (02) :132-141
[6]   TextFlow: Towards Better Understanding of Evolving Topics in Text [J].
Cui, Weiwei ;
Liu, Shixia ;
Tan, Li ;
Shi, Conglei ;
Song, Yangqiu ;
Gao, Zekai J. ;
Tong, Xin ;
Qu, Huamin .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2011, 17 (12) :2412-2421
[7]   Attention Flows: Analyzing and Comparing Attention Mechanisms in Language Models [J].
DeRose, Joseph F. ;
Wang, Jiayao ;
Berger, Matthew .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2021, 27 (02) :1160-1170
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]   Interactive Attention Model Explorer for Natural Language Processing Tasks with Unbalanced Data Sizes [J].
Dong, Zhihang ;
Wu, Tongshuang ;
Song, Sicheng ;
Zhang, Mingrui .
2020 IEEE PACIFIC VISUALIZATION SYMPOSIUM (PACIFICVIS), 2020, :46-50
[10]  
Dosilovic FK, 2018, 2018 41ST INTERNATIONAL CONVENTION ON INFORMATION AND COMMUNICATION TECHNOLOGY, ELECTRONICS AND MICROELECTRONICS (MIPRO), P210, DOI 10.23919/MIPRO.2018.8400040