An Introductory Survey on Attention Mechanisms in NLP Problems

被引:191
作者
Hu, Dichao [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
来源
INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2 | 2020年 / 1038卷
关键词
Natural language processing; Attention; Deep learning;
D O I
10.1007/978-3-030-29513-4_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
First derived from human intuition, later adapted to machine translation for automatic token alignment, attention mechanism, a simple method that can be used for encoding sequence data based on the importance score each element is assigned, has been widely applied to and attained significant improvement in various tasks in natural language processing, including sentiment classification, text summarization, question answering, dependency parsing, etc. In this paper, we survey through recent works and conduct an introductory summary of the attention mechanism in different NLP problems, aiming to provide our readers with basic knowledge on this widely used method, discuss its different variants for different tasks, explore its association with other techniques in machine learning, and examine methods for evaluating its performance.
引用
收藏
页码:432 / 448
页数:17
相关论文
共 33 条
[1]  
[Anonymous], 2018, P C N AM CHAPT ASS C, DOI DOI 10.18653/V1/N18-2074
[2]  
[Anonymous], 2018, P C EMP METH NAT LAN
[3]  
[Anonymous], 2016, ARXIV160601549
[4]  
[Anonymous], P INT C NEUR INF PRO
[5]  
[Anonymous], ARXIV150804025, DOI DOI 10.18653/V1/2021.FNDINGSEMNLP.101
[6]  
[Anonymous], 2017, INT C LEARNING REPRE
[7]  
[Anonymous], 2018, ARXIV180507037
[8]  
Bahdanau D, 2014, 3 INT C LEARN REPR
[9]  
Cho K., 2014, P SSST 8 8 WORKSH SY, P103
[10]  
Chung J., 2014, NIPS 2014 WORKSHOP D