Multi-view learning with Universum

被引:26
作者
Wang, Zhe [1 ]
Zhu, Yujin [1 ]
Liu, Wenwen [1 ]
Chen, Zhihua [1 ]
Gao, Daqi [1 ]
机构
[1] E China Univ Sci & Technol, Dept Comp Sci & Engn, Shanghai 200237, Peoples R China
关键词
Multi-view learning; Universum learning; Regularization learning; Rademacher complexity; Pattern classification; SUPPORT VECTOR MACHINE;
D O I
10.1016/j.knosys.2014.07.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The traditional Multi-view Learning (MVL) studies how to process patterns with multiple information sources. In practice, the MVL is proven to have a significant advantage over the Single-view Learning (SVL). But in most real-world cases, there are only single-source patterns to be dealt with and the existing MVL is unable to be directly applied. In order to solve this problem, an alternative MVL technique was developed for the single-source patterns through reshaping the original vector representation of the single-source patterns into multiple matrix representations in our previous work. Doing so can effectively bring an improved classification performance. This paper aims to generalize the previous MVL through taking advantage of the Universum examples which do not belong to either class of the classification problem. The newly-proposed generalization can not only inherit the advantage of the previous MVL, but also get a prior domain knowledge of the whole data distribution. To our knowledge, it introduces the Universum technique into the MVL for the first time. In the implementation, our previous MVL named MultiV-MHKS is selected as the learning paradigm and incorporate MultiV-MHKS with the Universum technique, which forms a more flexible MVL with the Universum called UMultiV-MHKS for short. The subsequent experiments validate that the proposed UMultiV-MHKS can effectively improve the classification performance over both the original MultiV-MHKS and some other state-of-the-art algorithms. Finally, it is demonstrated that the UMultiV-MHKS can get a tighter generalization risk bound in terms of the Rademacher complexity. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:376 / 391
页数:16
相关论文
共 41 条
[1]  
Abney S, 2002, 40TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, P360
[2]  
[Anonymous], 2007, LEARN DATA CONCEPTS, DOI DOI 10.1002/9780470140529.CH4.[38]L
[3]  
[Anonymous], 2008, P 2008 SIAM INT C DA
[4]  
[Anonymous], 1997, MACHINE LEARNING, MCGRAW-HILL SCIENCE/ENGINEERING/MATH
[5]  
[Anonymous], 2008, NEURAL INFORM PROCES
[6]  
[Anonymous], 1992, BREAKTHROUGHS STAT M, DOI DOI 10.1007/978-1-4612-4380-9_14
[7]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690
[8]  
Bin-Bin Gao, 2012, 2012 5th International Conference on BioMedical Engineering and Informatics (BMEI), P1265, DOI 10.1109/BMEI.2012.6513173
[9]  
Blum A., 1998, Proceedings of the Eleventh Annual Conference on Computational Learning Theory, P92, DOI 10.1145/279943.279962
[10]  
Breiman L, 1996, ANN STAT, V24, P2350