Bayesian multitask learning regression for heterogeneous patient cohorts

被引:7
作者
Goncalves A. [1 ]
Ray P. [1 ]
Soper B. [1 ]
Widemann D. [1 ]
Nygård M. [2 ]
Nygård J.F. [2 ]
Sales A.P. [1 ]
机构
[1] Lawrence Livermore National Laboratory, Livermore, CA
[2] Cancer Registry of Norway, Oslo
来源
Journal of Biomedical Informatics: X | 2019年 / 4卷
关键词
Alzheimer's disease progression; Bayesian modeling; Biomedical application; Multitask learning; Structured learning; Uncertainty quantification;
D O I
10.1016/j.yjbinx.2019.100059
中图分类号
学科分类号
摘要
Multitask learning (MTL) leverages commonalities across related tasks with the aim of improving individual task performance. A key modeling choice in designing MTL models is the structure of the tasks’ relatedness, which may not be known. Here we propose a Bayesian multitask learning model that is able to infer the task relationship structure directly from the data. We present two variations of the model in terms of a priori information of task relatedness. First, a diffuse Wishart prior is placed on a task precision matrix so that all tasks are assumed to be equally related a priori. Second, a Bayesian graphical LASSO prior is used on the task precision matrix to impose sparsity in the task relatedness. Motivated by machine learning applications in the biomedical domain, we emphasize interpretability and uncertainty quantification in our models. To encourage model interpretability, linear mappings from the shared input spaces to task-dependent output spaces are used. To encourage uncertainty quantification, conjugate priors are used so that full posterior inference is possible. Using synthetic data, we show that our model is able to recover the underlying task relationships as well as features jointly relevant for all tasks. We demonstrate the utility of our model on three distinct biomedical applications: Alzheimer's disease progression, Parkinson's disease assessment, and cervical cancer screening compliance. We show that our model outperforms Single Task (STL) models in terms of predictive performance, and performs better than existing MTL methods for the majority of the scenarios. © 2019 The Authors
引用
收藏
相关论文
共 46 条
[1]  
Caruana R., Multitask learning, Machine Learn., 28, 1, pp. 41-75, (1997)
[2]  
Widmer C., Ratsch G., Multitask learning in computational biology, ICML Workshop on Unsupervised and Transfer Learning, pp. 207-216, (2012)
[3]  
Wang X., Zhang C., Zhang Z., Boosted multi-task learning for face verification with applications to web image and video search, IEEE Conference on Computer Vision and Pattern Recognition, pp. 142-149, (2009)
[4]  
Ray P., Zheng L., Lucas J., Carin L., Bayesian joint analysis of heterogeneous genomics data, Bioinformatics, 30, 10, pp. 1370-1376, (2014)
[5]  
Glorot X., Bordes A., Bengio Y., Domain adaptation for large-scale sentiment classification: A deep learning approach, International Conference on Machine Learning (ICML), pp. 513-520, (2011)
[6]  
Zhang Y., Yang Q., pp. 1-20, (2017)
[7]  
Bonilla E.V., Chai K.M., Williams C., Multi-task Gaussian process prediction, Advances in Neural Information Processing Systems, pp. 153-160, (2008)
[8]  
Ruder S., pp. 1-14, (2017)
[9]  
Jacob L., Vert J.-P., Bach F.R., Clustered multi-task learning: A convex formulation, Advances in Neural Information Processing Systems, pp. 745-752, (2009)
[10]  
Xue Y., Liao X., Carin L., Krishnapuram B., Multi-task learning for classification with Dirichlet process priors, J. Mach. Learn. Res., 8, Jan, pp. 35-63, (2007)