Correlational Neural Networks

被引:78
作者
Chandar, Sarath [1 ]
Khapra, Mitesh M. [2 ]
Larochelle, Hugo [3 ]
Ravindran, Balaraman [4 ]
机构
[1] Univ Montreal, Montreal, PQ H3T 1J4, Canada
[2] IBM Res India, Bangalore 560077, Karnataka, India
[3] Univ Sherbrooke, Sherbrooke, PQ J1K 2R1, Canada
[4] Indian Inst Technol, Madras 600036, Tamil Nadu, India
关键词
CANONICAL CORRELATION-ANALYSIS;
D O I
10.1162/NECO_a_00801
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Common representation learning (CRL), wherein different descriptions (or views) of the data are embedded in a common subspace, has been receiving a lot of attention recently. Two popular paradigms here are canonical correlation analysis (CCA)-based approaches and autoencoder (AE)-based approaches. CCA-based approaches learn a joint representation by maximizing correlation of the views when projected to the common subspace. AE-based methods learn a common representation by minimizing the error of reconstructing the two views. Each of these approaches has its own advantages and disadvantages. For example, while CCA-based approaches outperform AE-based approaches for the task of transfer learning, they are not as scalable as the latter. In this work, we propose an AE-based approach, correlational neural network (CorrNet), that explicitly maximizes correlation among the views when projected to the common subspace. Through a series of experiments, we demonstrate that the proposed CorrNet is better than AE and CCA with respect to its ability to learn correlated common representations. We employ CorrNet for several cross-language tasks and show that the representations learned using it perform better than the ones learned using other state-of-the-art approaches.
引用
收藏
页码:257 / 285
页数:29
相关论文
共 38 条
[1]  
Akaho S., 2001, INT M PSYCH SOC, P263
[2]  
[Anonymous], 2015, CORRABS150203167
[3]  
[Anonymous], P PYTH SCI COMP C
[4]  
[Anonymous], P INT C MACH LEARN
[5]  
[Anonymous], 2013, P INT C MACH LEARN
[6]  
[Anonymous], P INT C LEARN REPR
[7]  
[Anonymous], 2011, Advances in Neural Information Processing Systems
[8]  
[Anonymous], 2009, NATURAL LANGUAGE PRO, DOI DOI 10.1007/S10579-010-9124-X
[9]  
Arora R., 2012, PROC MACHINE LEARNIN, P34
[10]  
Chandar S., 2013, NIPS DEEP LEARN WORK