RAGAN: Regression Attention Generative Adversarial Networks

被引:7
作者
Jiang X. [1 ]
Ge Z. [1 ,2 ]
机构
[1] Zhejiang University, State Key Laboratory of Industrial Control Technology, College of Control Science and Engineering, Hangzhou
[2] Peng Cheng Laboratory, Shenzhen
来源
IEEE Transactions on Artificial Intelligence | 2023年 / 4卷 / 06期
基金
中国国家自然科学基金;
关键词
Data augmentation; evaluation indicator; generative adversarial networks (GANs); regression attention; regression model;
D O I
10.1109/TAI.2022.3209956
中图分类号
学科分类号
摘要
Despite surrounding by Big Data, we still need to learn from insufficient data in many scenarios. Building an accurate regression model for a small amount of data is a pretty tricky and exciting problem. At present, it is a promising solution to augment limited real data by generating data through generative adversarial networks (GANs). However, when GAN is used to generate labeled data in regression modeling, it lacks attention to the relationship between independent and dependent variables, resulting in poor performance of regression modeling. This article proposes a novel regression attention GAN (RA-GAN) for augmented regression modeling. Regression attention mechanisms are introduced into network parameters learning of both generator and discriminator in RA-GAN to establish a known relationship between variables. This makes RA-GAN restore the regression information during data generation. In addition, an indicator called cross regression score is designed to describe the quality of the generated data before augmented regression modeling, effectively evaluating data augmentation performance in advance. The effectiveness and superiority of the proposed methods are verified in an actual industrial soft-sensing case and a diabetes prediction case through data augmentation regression applications. © 2020 IEEE.
引用
收藏
页码:1549 / 1563
页数:14
相关论文
共 31 条
[21]  
Salimans T., Goodfellow I., Zaremba W., Cheung V., Radford A., Chen X., Improved techniques for training GANs, Proc. Int.Conf.Neural Inf. Process. Syst., pp. 2234-2242, (2016)
[22]  
Che T., Li Y., Jacob A.P., Bengio Y., Li W., Mode regularized generative adversarial networks, Proc. 5th Int. Conf. Learn. Representations, (2017)
[23]  
Heusel M., Ramsauer H., Unterthiner T., Nessler B., Hochreiter S., GANs trained by a two time-scale update rule converge to a local Nash equilibrium, Proc. Int. Conf. Neural Inf. Process. Syst., pp. 6626-6637, (2017)
[24]  
Arjovsky M., Chintala S., Bottou L., Wasserstein generative adversarial networks, Proc. 34th Int. Conf. Mach. Learn., 70, pp. 214-223, (2017)
[25]  
Gulrajani I., Ahmed F., Arjovsky M., Dumoulin V., Courville A.C., Improved training of Wasserstein GANs, Proc. Int. Conf. Neural Inf. Process. Syst., pp. 5767-5777, (2017)
[26]  
Lerner B., Guterman H., Aladjem M., Dinstein I.H., A comparative study of neural network based feature extraction paradigms, Pattern Recognit. Lett., 20, 1, pp. 7-14, (1999)
[27]  
Mao J., Jain A.K., Artificial neural networks for feature extraction and multivariate data projection, IEEE Trans. Neural Netw., 6, 2, pp. 296-317, (1995)
[28]  
Geisser S., Eddy W.F., A predictive approach to model selection, J. Amer. Stat. Assoc., 74, 365, pp. 153-160, (1979)
[29]  
Altman N.S., An introduction to kernel and nearest-neighbor nonparametric regression, Amer. Statistician, 46, 3, pp. 175-185, (1992)
[30]  
Efron B., Hastie T., Johnstone I., Tibshirani R., Least angle regression, Ann. Statist., 32, 2, pp. 407-499, (2004)