Deep Active Learning for Text Classification

被引:7
作者
An, Bang [1 ]
Wu, Wenjun [1 ]
Han, Huimin [1 ]
机构
[1] Beihang Univ, 37 XueYuan Rd, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING (ICVISP 2018) | 2018年
关键词
Active Learning; Deep Learning; Machine Learning; Artificial Intelligence; Text Classification;
D O I
10.1145/3271553.3271578
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In recent years, Active Learning (AL) has been applied in the domain of text classification successfully. However, traditional methods need researchers to pay attention to feature extraction of datasets and different features will influence the final accuracy seriously. In this paper, we propose a new method that uses Recurrent Neutral Network (RNN) as the acquisition function in Active Learning called Deep Active Learning (DAL). For DAL, there is no need to consider how to extract features because RNN can use its internal state to process sequences of inputs. We have proved that DAL can achieve the accuracy that cannot be reached by traditional Active Learning methods when dealing with text classification. What's more, DAL can decrease the need of the great number of labeled instances for Deep Learning (DL). At the same time, we design a strategy to distribute label work to different workers. We have proved by using a proper batch size of instance, we can save much time but not decrease the model's accuracy. Based on this, we provide batch of instances for different workers and the size of batch is determined by worker's ability and scale of dataset, meanwhile, it can be updated with the performance of the workers.
引用
收藏
页数:6
相关论文
共 19 条
[1]  
Aggarwal Charu C, 2012, Mining text data, P163, DOI [DOI 10.1007/978-1-4614-3223-46, DOI 10.1007/978-1-4614-3223-4, 10.1007/978-1-4614-3223-4]
[2]  
Chen Y., 2013, ICML 13 P 30 INT C I
[3]  
Gal Y., 2016, THESIS
[4]  
Gal Y, 2016, ADV NEUR IN, V29
[5]  
Hassan S., 2011, IEEE 14 INT MULT C
[6]  
Hoi S., 2006, ICML 06 P 23 INT C M
[7]  
Huang K., 2016, 2016 IEEE 16 INT C D
[8]  
Kim Y., 2014, P 2014 C EMP METH NA, DOI [10.3115/v1/D14-1181, DOI 10.3115/V1/D14-1181]
[9]  
Lai SW, 2015, AAAI CONF ARTIF INTE, P2267
[10]  
Lilleberg J., 2015, IEEE 14 INT C COGN I