Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression

被引:0
|
作者
Karmitsa, Napsu [1 ]
Taheri, Sona [2 ]
Joki, Kaisa [3 ]
Paasivirta, Pauliina [4 ]
Bagirov, Adil M. [5 ]
Makela, Marko M. [3 ]
机构
[1] Univ Turku, Dept Comp, FI-20014 Turku, Finland
[2] RMIT Univ, Sch Sci, Melbourne 3000, Australia
[3] Univ Turku, Dept Math & Stat, FI-20014 Turku, Finland
[4] Siili Solut Oyj, FI-60100 Seinajoki, Finland
[5] Federat Univ Australia, Ctr Smart Analyt, Ballarat 3350, Australia
基金
芬兰科学院; 澳大利亚研究理事会;
关键词
machine learning; regression analysis; neural networks; L1-loss function; nonsmooth optimization; PERFORMANCE; REPRESENTATIONS; PARAMETERS; MACHINE;
D O I
10.3390/a16090444
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a new nonsmooth optimization-based algorithm for solving large-scale regression problems is introduced. The regression problem is modeled as fully-connected feedforward neural networks with one hidden layer, piecewise linear activation, and the L1-loss functions. A modified version of the limited memory bundle method is applied to minimize this nonsmooth objective. In addition, a novel constructive approach for automated determination of the proper number of hidden nodes is developed. Finally, large real-world data sets are used to evaluate the proposed algorithm and to compare it with some state-of-the-art neural network algorithms for regression. The results demonstrate the superiority of the proposed algorithm as a predictive tool in most data sets used in numerical experiments.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] A SUPERCONDUCTING NEURAL CELL SUITABLE FOR LARGE-SCALE NEURAL NETWORKS
    HIDAKA, M
    AKERS, LA
    APPLIED SUPERCONDUCTIVITY, 1993, 1 (10-12) : 1907 - 1919
  • [42] A Population-Based Hybrid Approach for Hyperparameter Optimization of Neural Networks
    Japa, Luis
    Serqueira, Marcello
    Mendonca, Israel
    Aritsugi, Masayoshi
    Bezerra, Eduardo
    Gonzalez, Pedro Henrique
    IEEE ACCESS, 2023, 11 : 50752 - 50768
  • [43] Scour modeling using deep neural networks based on hyperparameter optimization
    Asim, Mohammed
    Rashid, Adnan
    Ahmad, Tanvir
    ICT EXPRESS, 2022, 8 (03): : 357 - 362
  • [44] Globally convergent limited memory bundle method for large-scale nonsmooth optimization
    Napsu Haarala
    Kaisa Miettinen
    Marko M. Mäkelä
    Mathematical Programming, 2007, 109 : 181 - 205
  • [45] A Modified Nonlinear Conjugate Gradient Algorithm for Large-Scale Nonsmooth Convex Optimization
    Woldu, Tsegay Giday
    Zhang, Haibin
    Zhang, Xin
    Fissuh, Yemane Hailu
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2020, 185 (01) : 223 - 238
  • [46] A Modified Nonlinear Conjugate Gradient Algorithm for Large-Scale Nonsmooth Convex Optimization
    Tsegay Giday Woldu
    Haibin Zhang
    Xin Zhang
    Yemane Hailu Fissuh
    Journal of Optimization Theory and Applications, 2020, 185 : 223 - 238
  • [47] Globally convergent limited memory bundle method for large-scale nonsmooth optimization
    Haarala, Napsu
    Miettinen, Kaisa
    Makela, Marko M.
    MATHEMATICAL PROGRAMMING, 2007, 109 (01) : 181 - 205
  • [48] Distributed and Cooperative Optimization-based Iterative Learning Control for Large-scale Building Temperature Regulation
    Peng, Cheng
    Sun, Liting
    Tomizuka, Masayoshi
    2017 IEEE INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2017, : 1606 - 1611
  • [49] A distributed optimization-based approach for hierarchical MPC of large-scale systems with coupled dynamics and constraints
    Minh Dang Doan
    Keviczky, Tamas
    De Schutter, Bart
    2011 50TH IEEE CONFERENCE ON DECISION AND CONTROL AND EUROPEAN CONTROL CONFERENCE (CDC-ECC), 2011, : 5236 - 5241
  • [50] Start Small, Think Big: On Hyperparameter Optimization for Large-Scale Knowledge Graph Embeddings
    Kochsiek, Adrian
    Niesel, Fritz
    Gemulla, Rainer
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT II, 2023, 13714 : 138 - 154