An Interpretable Knowledge Representation Framework for Natural Language Processing with Cross-Domain Application

被引:6
作者
Bhattarai, Bimal [1 ]
Granmo, Ole-Christoffer [1 ]
Jiao, Lei [1 ]
机构
[1] Univ Agder, Ctr AI Res, Grimstad, Norway
来源
ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT I | 2023年 / 13980卷
关键词
Natural language processing (NLP); Tsetlin machine (TM); Propositional logic; Knowledge representation; Domain adaptation; Interpretable representation;
D O I
10.1007/978-3-031-28244-7_11
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Data representation plays a crucial role in natural language processing (NLP), forming the foundation for most NLP tasks. Indeed, NLP performance highly depends upon the effectiveness of the preprocessing pipeline that builds the data representation. Many representation learning frameworks, such asWord2Vec, encode input data based on local contextual information that interconnects words. Such approaches can be computationally intensive, and their encoding is hard to explain. We here propose an interpretable representation learning framework utilizing Tsetlin Machine (TM). The TM is an interpretable logic-based algorithm that has exhibited competitive performance in numerous NLP tasks. We employ the TM clauses to build a sparse propositional (boolean) representation of natural language text. Each clause is a class-specific propositional rule that links words semantically and contextually. Through visualization, we illustrate how the resulting data representation provides semantically more distinct features, better separating the underlying classes. As a result, the following classification task becomes less demanding, benefiting simple machine learning classifiers such as Support Vector Machine (SVM). We evaluate our approach using six NLP classification tasks and twelve domain adaptation tasks. Our main finding is that the accuracy of our proposed technique significantly outperforms the vanilla TM, approaching the competitive accuracy of deep neural network (DNN) baselines. Furthermore, we present a case study showing how the representations derived from our framework are interpretable. (We use an asynchronous and parallel version of Tsetlin Machine: available at https://github.com/cair/PyTsetlinMachineCUDA).
引用
收藏
页码:167 / 181
页数:15
相关论文
共 51 条
[1]   The regression Tsetlin machine: a novel approach to interpretable nonlinear regression [J].
Abeyrathna, K. Darshana ;
Granmo, Ole-Christoffer ;
Zhang, Xuan ;
Jiao, Lei ;
Goodwin, Morten .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2020, 378 (2164)
[2]  
Abeyrathna KD, 2021, PR MACH LEARN RES, V139
[3]  
[Anonymous], 2004, Proceedings of the ACL
[4]  
[Anonymous], 2005, P HUM LANG TECHN C C, DOI DOI 10.3115/1220575.1220619
[5]   On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? [J].
Bender, Emily M. ;
Gebru, Timnit ;
McMillan-Major, Angelina ;
Shmitchell, Shmargaret .
PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, :610-623
[6]  
Bengio Y, 2001, ADV NEUR IN, V13, P932
[7]  
Bengio Y., 2008, SCHOLARPEDIA J, V3, P3881, DOI DOI 10.4249/SCHOLARPEDIA.3881
[8]   Using the Tsetlin Machine to Learn Human-Interpretable Rules for High-Accuracy Text Categorization With Medical Applications [J].
Berge, Geir Thore ;
Granmo, Ole-Christoffer ;
Tveit, Tor Oddbjorn ;
Goodwin, Morten ;
Jiao, Lei ;
Matheussen, Bernt Viggo .
IEEE ACCESS, 2019, 7 :115134-115146
[9]  
Bhattarai B, 2022, LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, P3761
[10]  
Bhattarai B, 2022, LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, P4894