Local ensemble learning from imbalanced and noisy data for word sense disambiguation

被引:18
作者
Krawczyk, Bartosz [1 ]
McInnes, Bridget T. [1 ]
机构
[1] Virginia Commonwealth Univ, Dept Comp Sci, Richmond, VA 23284 USA
关键词
Machine learning; Natural language processing; Imbalanced classification; Multi-class imbalance; Ensemble learning; One-class classification; Class label noise; Word sense disambiguation; SAMPLING APPROACH; CLASSIFICATION; ALGORITHMS;
D O I
10.1016/j.patcog.2017.10.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Natural Language Processing plays a key role in man-machine interactions, allowing computers to understand and analyze human language. One of its more challenging sub-domains is word sense disambiguation, the task of automatically identifying the intended sense (or concept) of an ambiguous word based on the context in which the word is used. This requires proper feature extraction to capture specific data properties and a dedicated machine learning solution to allow for the accurate labeling of the appropriate sense. However, the pattern classification problem posed here is highly challenging, as we must deal with high-dimensional and multi-class imbalanced data that additionally may be corrupted with class label noise. To address these issues, we propose a local ensemble learning solution. It uses a one-class decomposition of the multi-class problem, assigning an ensemble of one-class classifiers to each of the distributions. The classifiers are trained on the basis of low-dimensional subsets of features and a kernel feature space transformation to obtain a more compact representation. Instance weighting is used to filter out potentially noisy instances and reduce overlapping among classes. Finally, a two-level classifier fusion technique is used to reconstruct the original multi-class problem. Our results show that the proposed learning approach displays robustness to both multi-class skewed distributions and class label noise, making it a useful tool for the considered task. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:103 / 119
页数:17
相关论文
共 57 条
[1]  
Agirre E, 2006, TEXT SPEECH LANG TEC, V33, P1, DOI 10.1007/978-1-4020-4809-8
[2]   An Efficient Over-sampling Approach Based on Mean Square Error Back-propagation for Dealing with the Multi-class Imbalance Problem [J].
Alejo, R. ;
Garcia, V. ;
Pacheco-Sanchez, J. H. .
NEURAL PROCESSING LETTERS, 2015, 42 (03) :603-617
[3]   One-Class versus Binary Classification: Which and When? [J].
Bellinger, Colin ;
Sharma, Shiven ;
Japkowicz, Nathalie .
2012 11TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2012), VOL 2, 2012, :102-106
[4]   Soft clustering using weighted one-class support vector machines [J].
Bicego, Manuele ;
Figueiredo, Mario A. T. .
PATTERN RECOGNITION, 2009, 42 (01) :27-32
[5]  
BRUCE R, 1994, 32ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, P139
[6]   ur-CAIM: improved CAIM discretization for unbalanced and balanced data [J].
Cano, Alberto ;
Nguyen, Dat T. ;
Ventura, Sebastian ;
Cios, Krzysztof J. .
SOFT COMPUTING, 2016, 20 (01) :173-188
[7]   Weighted Data Gravitation Classification for Standard and Imbalanced Data [J].
Cano, Alberto ;
Zafra, Amelia ;
Ventura, Sebastian .
IEEE TRANSACTIONS ON CYBERNETICS, 2013, 43 (06) :1672-1687
[8]   Near-Bayesian Support Vector Machines for imbalanced data classification with equal or unequal misclassification costs [J].
Datta, Shounak ;
Das, Swagatam .
NEURAL NETWORKS, 2015, 70 :39-52
[9]  
Dunning T., 1993, Computational Linguistics, V19, P61
[10]  
Edmonds Philip, 2001, P SENSEVAL2 2 INT WO, P1