Inductive transfer for learning Bayesian networks

被引:0
作者
Roger Luis
L. Enrique Sucar
Eduardo F. Morales
机构
[1] Instituto Nacional de Astrofísica,
[2] Óptica y Electrónica (INAOE),undefined
来源
Machine Learning | 2010年 / 79卷
关键词
Inductive transfer; Bayesian networks; Structure learning; Parameter learning;
D O I
暂无
中图分类号
学科分类号
摘要
In several domains it is common to have data from different, but closely related problems. For instance, in manufacturing, many products follow the same industrial process but with different conditions; or in industrial diagnosis, where there is equipment with similar specifications. In these cases it is common to have plenty of data for some scenarios but very little for others. In order to learn accurate models for rare cases, it is desirable to use data and knowledge from similar cases; a technique known as transfer learning. In this paper we propose an inductive transfer learning method for Bayesian networks, that considers both structure and parameter learning. For structure learning we use conditional independence tests, by combining measures from the target task with those obtained from one or more auxiliary tasks, using a novel weighted sum of the conditional independence measures. For parameter learning, we propose two variants of the linear pool for probability aggregation, combining the probability estimates from the target task with those from the auxiliary tasks. To validate our approach, we used three Bayesian networks models that are commonly used for evaluating learning techniques, and generated variants of each model by changing the structure as well as the parameters. We then learned one of the variants with a small dataset and combined it with information from the other variants. The experimental results show a significant improvement in terms of structure and parameters when we transfer knowledge from similar tasks. We also evaluated the method with real-world data from a manufacturing process considering several products, obtaining an improvement in terms of log-likelihood between the data and the model when we do transfer learning from related products.
引用
收藏
页码:227 / 255
页数:28
相关论文
共 20 条
[1]  
Baxter J.(1997)A Bayesian/information theoretic model of learning to learn via multiple task sampling Machine Learning 28 7-39
[2]  
Binder J.(1997)Adaptive probabilistic networks with hidden variables Machine Learning 29 213-244
[3]  
Koller D.(1997)Multitask learning Machine Learning 28 41-75
[4]  
Russell S.(1996)Aggregate functions over probabilisitic data Informing Science 88 15-45
[5]  
Kanazawa K.(1996)Evaluating aggregate operations over imprecise data IEEE Transactions on Knowledge and Data Engineering 8 273-284
[6]  
Caruana R.(1992)A Bayesian method for the induction of probabilistic networks from data Machine Learning 9 309-347
[7]  
Chang C.-S.(1986)Combining probability distributions: a critique and an annotated bibliography Statistical Science 1 114-148
[8]  
Chen A. L. P.(1994)Learning Bayesian belief networks: an approach based on the MDL principle Computational Intelligence 10 269-293
[9]  
Chen A. L. P.(2008)Inductive transfer with context-sensitive neural networks Machine Learning 73 313-336
[10]  
Chiu J.-S.(undefined)undefined undefined undefined undefined-undefined