Nonlinear Regression via Deep Negative Correlation Learning

被引:50
作者
Zhang, Le [1 ]
Shi, Zenglin [2 ]
Cheng, Ming-Ming [3 ]
Liu, Yun [3 ]
Bian, Jia-Wang [4 ]
Zhou, Joey Tianyi [1 ]
Zheng, Guoyan [5 ]
Zeng, Zeng [1 ]
机构
[1] ASTAR, Singapore 138632, Singapore
[2] Univ Amsterdam, NL-1012 WX Amsterdam, Netherlands
[3] Nankai Univ, Coll Comp Sci, TKLNDST, Nankai 300071, Peoples R China
[4] Univ Adelaide, Sch Comp Sci, Adelaide, SA 5005, Australia
[5] Shanghai Jiao Tong Univ, Sch Biomed Engn, Shanghai 200240, Peoples R China
关键词
Task analysis; Estimation; Training; Correlation; Computational modeling; Deep learning; Computer vision; deep regression; negative correlation learning; convolutional neural network; IMAGE QUALITY ASSESSMENT; HUMAN AGE ESTIMATION; SUPERRESOLUTION; ENSEMBLES; CASCADE;
D O I
10.1109/TPAMI.2019.2943860
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonlinear regression has been extensively employed in many computer vision problems (e.g., crowd counting, age estimation, affective computing). Under the umbrella of deep learning, two common solutions exist i) transforming nonlinear regression to a robust loss function which is jointly optimizable with the deep convolutional network, and ii) utilizing ensemble of deep networks. Although some improved performance is achieved, the former may be lacking due to the intrinsic limitation of choosing a single hypothesis and the latter may suffer from much larger computational complexity. To cope with those issues, we propose to regress via an efficient "divide and conquer" manner. The core of our approach is the generalization of negative correlation learning that has been shown, both theoretically and empirically, to work well for non-deep regression problems. Without extra parameters, the proposed method controls the bias-variance-covariance trade-off systematically and usually yields a deep regression ensemble where each base model is both "accurate" and "diversified." Moreover, we show that each sub-problem in the proposed method has less Rademacher Complexity and thus is easier to optimize. Extensive experiments on several diverse and challenging tasks including crowd counting, personality analysis, age estimation, and image super-resolution demonstrate the superiority over challenging baselines as well as the versatility of the proposed method. The source code and trained models are available on our project page: https://mmcheng.net/dncl/.
引用
收藏
页码:982 / 998
页数:17
相关论文
共 50 条
  • [41] Relieving Coefficient Learning in Genetic Programming for Symbolic Regression via Correlation and Linear Scaling
    Chen, Qi
    Xue, Bing
    Banzhaf, Wolfgang
    Zhang, Mengjie
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, GECCO 2023, 2023, : 420 - 437
  • [42] Underwater Image Co-Enhancement With Correlation Feature Matching and Joint Learning
    Qi, Qi
    Zhang, Yongchang
    Tian, Fei
    Wu, Q. M. Jonathan
    Li, Kunqian
    Luan, Xin
    Song, Dalei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (03) : 1133 - 1147
  • [43] FAST APPROXIMATION OF NON-NEGATIVE SPARSE RECOVERY VIA DEEP LEARNING
    Xie, Youye
    Wang, Zifan
    Pei, Weiping
    Tang, Gongguo
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 2921 - 2925
  • [44] Reconstructing Unsteady Flow Data From Representative Streamlines via Diffusion and Deep-Learning-Based Denoising
    Gu, Pengfei
    Han, Jun
    Chen, Danny Z.
    Wang, Chaoli
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2021, 41 (06) : 111 - 121
  • [45] Multilinear Kernel Regression and Imputation via Manifold Learning
    Nguyen, Duc Thien
    Slavakis, Konstantinos
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2024, 5 : 1073 - 1088
  • [46] EdgePro: Edge Deep Learning Model Protection via Neuron Authorization
    Chen, Jinyin
    Zheng, Haibin
    Liu, Tao
    Liu, Jiawei
    Cheng, Yao
    Zhang, Xuhong
    Ji, Shouling
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (05) : 4967 - 4981
  • [47] The Negative BER Loss Function for Deep Learning Decoders
    Dong, Rui
    Lu, Fang
    Dong, Yan
    Yan, Haotian
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (08) : 1824 - 1828
  • [48] A Comprehensive Analysis of Deep Regression
    Lathuiliere, Stephane
    Mesejo, Pablo
    Alameda-Pineda, Xavier
    Horaud, Radu
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (09) : 2065 - 2081
  • [49] Deep Metric Learning for Crowdedness Regression
    Wang, Qi
    Wan, Jia
    Yuan, Yuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (10) : 2633 - 2643
  • [50] Regression Fuzzing for Deep Learning Systems
    You, Hanmo
    Wang, Zan
    Chen, Junjie
    Liu, Shuang
    Li, Shuochuan
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 82 - 94