Representation learning with deep sparse auto-encoder for multi-task learning

被引:13
作者
Zhu, Yi [1 ,2 ,3 ]
Wu, Xindong [2 ,3 ]
Qiang, Jipeng [1 ]
Hu, Xuegang [2 ,3 ]
Zhang, Yuhong [2 ,3 ]
Li, Peipei [2 ,3 ]
机构
[1] Yangzhou Univ, Sch Informat Engn, Yangzhou, Peoples R China
[2] Hefei Univ Technol, Key Lab Knowledge Engn Big Data, Minist Educ China, Hefei, Peoples R China
[3] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep sparse auto-encoder; Multi-task learning; RICA; Labeled and unlabeled data; SUPPORT VECTOR MACHINES; FEATURE-SELECTION; REGULARIZATION; KNOWLEDGE;
D O I
10.1016/j.patcog.2022.108742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We demonstrate an effective framework to achieve a better performance based on Deep Sparse auto encoder for Multi-task Learning, called DSML for short. To learn the reconstructed and higher-level features on cross-domain instances for multiple tasks, we combine the labeled and unlabeled data from all tasks to reconstruct the feature representations. Furthermore, we propose the model of Stacked Reconstruction Independence Component Analysis (SRICA for short) for the optimization of feature representations with a large amount of unlabeled data, which can effectively address the redundancy of image data. Our proposed SRICA model is developed from RICA and is based on deep sparse auto-encoder. In addition, we adopt a Semi-Supervised Learning method (SSL for short) based on model parameter regularization to build a unified model for multi-task learning. There are several advantages in our proposed framework as follows: 1) The proposed SRICA makes full use of a large amount of unlabeled data from all tasks. It is used to pursue an optimal sparsity feature representation, which can overcome the over fitting problem effectively. 2) The deep architecture used in our SRICA model is applied for higher-level and better representation learning, which is designed to train on patches for sphering the input data. 3) Training parameters in our proposed framework has lower computational cost compared to other common deep learning methods such as stacked denoising auto-encoders. Extensive experiments on several real image datasets demonstrate our proposed framework outperforms the state-of-the-art methods.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
[21]   Minimum Description Length Penalization for Group and Multi-Task Sparse Learning [J].
Dhillon, Paramveer S. ;
Foster, Dean P. ;
Ungar, Lyle H. .
JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 :525-564
[22]   A multi-task deep learning framework for forecasting sparse demand of demand responsive transit [J].
Lee, Jaehyung ;
Choi, Yoonseo ;
Kim, Jinhee .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 250
[23]   Multi-task Representation Learning for Travel Time Estimation [J].
Li, Yaguang ;
Fu, Kun ;
Wang, Zheng ;
Shahabi, Cyrus ;
Ye, Jieping ;
Liu, Yan .
KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, :1695-1704
[24]   Enhanced representation and multi-task learning for image annotation [J].
Binder, Alexander ;
Samek, Wojciech ;
Mueller, Klaus-Robert ;
Kawanabe, Motoaki .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2013, 117 (05) :466-478
[25]   Multi-task Learning via Non-sparse Multiple Kernel Learning [J].
Samek, Wojciech ;
Binder, Alexander ;
Kawanabe, Motoaki .
COMPUTER ANALYSIS OF IMAGES AND PATTERNS: 14TH INTERNATIONAL CONFERENCE, CAIP 2011, PT I, 2011, 6854 :335-342
[26]   Dermoscopic attributes classification using deep learning and multi-task learning [J].
Saitov, Irek ;
Polevaya, Tatyana ;
Filchenkov, Andrey .
9TH INTERNATIONAL YOUNG SCIENTISTS CONFERENCE IN COMPUTATIONAL SCIENCE, YSC2020, 2020, 178 :328-336
[27]   Multi-task Sparse Regression Metric Learning for Heterogeneous Classification [J].
Wu, Haotian ;
Zhou, Bin ;
Zhu, Pengfei ;
Hu, Qinghua ;
Shi, Hong .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II, 2019, 11728 :543-553
[28]   Accelerated Gradient Method for Multi-Task Sparse Learning Problem [J].
Chen, Xi ;
Pan, Weike ;
Kwok, James T. ;
Carbonell, Jaime G. .
2009 9TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, 2009, :746-+
[29]   Multi-task Sparse Structure Learning with Gaussian Copula Models [J].
Goncalves, Andre R. ;
Von Zuben, Fernando J. ;
Banerjee, Arindam .
JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17 :1-30
[30]   Multi -task learning using variational auto -encoder for sentiment classification [J].
Lu, Guangquan ;
Zhao, Xishun ;
Yin, Jian ;
Yang, Weiwei ;
Li, Bo .
PATTERN RECOGNITION LETTERS, 2020, 132 :115-122