Extended MDL principle for feature-based inductive transfer learning

被引:0
作者
Hao Shao
Bin Tong
Einoshin Suzuki
机构
[1] Kyushu University,Graduate School of Systems Life Sciences
[2] ISEE,Department of Informatics
[3] Kyushu University,undefined
来源
Knowledge and Information Systems | 2013年 / 35卷
关键词
Inductive transfer learning; Minimum description length principle; Feature-based transfer;
D O I
暂无
中图分类号
学科分类号
摘要
Transfer learning provides a solution in real applications of how to learn a target task where a large amount of auxiliary data from source domains are given. Despite numerous research studies on this topic, few of them have a solid theoretical framework and are parameter-free. In this paper, we propose an Extended Minimum Description Length Principle (EMDLP) for feature-based inductive transfer learning, in which both the source and the target data sets contain class labels and relevant features are transferred from the source domain to the target one. Unlike conventional methods, our encoding measure is based on a theoretical background and has no parameter. To obtain useful features to be used in the target task, we design an enhanced encoding length by adopting a code book that stores useful information obtained from the source task. With the code book that builds connections between the source and the target tasks, our EMDLP is able to evaluate the inferiority of the results of transfer learning with the add sum of the code lengths of five components: those of the corresponding two hypotheses, the two data sets with the help of the hypotheses, and the set of the transferred features. The proposed method inherits the nice property of the MDLP that elaborately evaluates the hypotheses and balances the simplicity of the hypotheses and the goodness-of-the-fit to the data. Extensive experiments using both synthetic and real data sets show that the proposed method provides a better performance in terms of the classification accuracy and is robust against noise.
引用
收藏
页码:365 / 389
页数:24
相关论文
共 33 条
[1]  
Adhikari A(2011)Study of select items in different data sources by grouping Knowl Inf Syst (KAIS) 27 23-43
[2]  
Ramachandrarao P(2010)Bulk construction of dynamic clustered metric trees Knowl Inf Syst (KAIS) 22 211-244
[3]  
Pedrycz W(2003)Task clustering and gating for bayesian multitask learning J Mach Learn Res 4 83-99
[4]  
Aronovich L(1997)Multitask learning Mach Learn 28 41-75
[5]  
Spiegler I(2006)Domain adaption for statistical classifiers J Artif Intell Res 26 101-126
[6]  
Bakker B(2008)Data discretization unification Knowl Inf Syst (KAIS) 19 1-29
[7]  
Heskes T(2001)PAC-Bayesian stochastic model selection Mach Learn J 51 5-21
[8]  
Caruana R(2008)A survey on transfer learning IEEE Trans Knowl Data Eng 22 1345-1359
[9]  
Daume H(1989)Inferring decision trees using the minimum description length principle Inf Comput 80 227-248
[10]  
Marcu D(1990)Learning logical definitions from relations Mach Learn 5 239-266