Neural Network with Optimal Neuron Activation Functions Based on Additive Gaussian Process Regression

被引:18
作者
Manzhos, Sergei [1 ]
Ihara, Manabu [1 ]
机构
[1] Tokyo Inst Technol, Sch Mat & Chem Technol, Meguro Ku, Tokyo 1528552, Japan
关键词
UNIVERSAL APPROXIMATION; NUMERICAL IMPLEMENTATION; SUPERPOSITION THEOREM; RS-HDMR; ENERGY; MACHINE; REPRESENTATION; SIMULATIONS; VARIABLES; BOUNDS;
D O I
10.1021/acs.jpca.3c02949
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Feed-forward neural networks (NNs) are a staple machine learning method widely used in many areas of science and technology, including physical chemistry, computational chemistry, and materials informatics. While even a single-hidden-layer NN is a universal approximator, its expressive power is limited by the use of simple neuron activation functions (such as sigmoid functions) that are typically the same for all neurons. More flexible neuron activation functions would allow the use of fewer neurons and layers and thereby save computational cost and improve expressive power. We show that additive Gaussian process regression (GPR) can be used to construct optimal neuron activation functions that are individual to each neuron. An approach is also introduced that avoids nonlinear fitting of neural network parameters by defining them with rules. The resulting method combines the advantage of robustness of a linear regression with the higher expressive power of an NN. We demonstrate the approach by fitting the potential energy surfaces of the water molecule and formaldehyde. Without requiring any nonlinear optimization, the additive-GPR-based approach outperforms a conventional NN in the high-accuracy regime, where a conventional NN suffers more from overfitting.
引用
收藏
页码:7823 / 7835
页数:13
相关论文
共 101 条
[51]   Optimization of hyperparameters of Gaussian process regression with the help of a low-order high-dimensional model representation: application to a potential energy surface [J].
Manzhos, Sergei ;
Ihara, Manabu .
JOURNAL OF MATHEMATICAL CHEMISTRY, 2023, 61 (01) :7-20
[52]   Easy representation of multivariate functions with low-dimensional terms via Gaussian process regression kernel design: applications to machine learning of potential energy surfaces and kinetic energy densities from sparse data [J].
Manzhos, Sergei ;
Sasaki, Eita ;
Ihara, Manabu .
MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01)
[53]   Neural Network Potential Energy Surfaces for Small Molecules and Reactions [J].
Manzhos, Sergei ;
Carrington, Tucker, Jr. .
CHEMICAL REVIEWS, 2021, 121 (16) :10187-10217
[54]   Machine learning for the solution of the Schrodinger equation [J].
Manzhos, Sergei .
MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2020, 1 (01)
[55]   Using an internal coordinate Gaussian basis and a space-fixed Cartesian coordinate kinetic energy operator to compute a vibrational spectrum with rectangular collocation [J].
Manzhos, Sergei ;
Carrington, Tucker .
JOURNAL OF CHEMICAL PHYSICS, 2016, 145 (22)
[56]   Neural network-based approaches for building high dimensional and quantum dynamics-friendly potential energy surfaces [J].
Manzhos, Sergei ;
Dawes, Richard ;
Carrington, Tucker .
INTERNATIONAL JOURNAL OF QUANTUM CHEMISTRY, 2015, 115 (16) :1012-1020
[57]   Fitting sparse multidimensional data with low-dimensional terms [J].
Manzhos, Sergei ;
Yamashita, Koichi ;
Carrington, Tucker, Jr. .
COMPUTER PHYSICS COMMUNICATIONS, 2009, 180 (10) :2002-2012
[58]   Using neural networks to represent potential surfaces as sums of products [J].
Manzhos, Sergei ;
Carrington, Tucker, Jr. .
JOURNAL OF CHEMICAL PHYSICS, 2006, 125 (19)
[59]   Machine-Learning-Assisted Discovery of High-Efficient Oxygen Evolution Electrocatalysts [J].
Mao, Xinnan ;
Wang, Lu ;
Li, Youyong .
JOURNAL OF PHYSICAL CHEMISTRY LETTERS, 2023, 14 (01) :170-177
[60]   Machine learning property prediction for organic photovoltaic devices [J].
Meftahi, Nastaran ;
Klymenko, Mykhailo ;
Christofferson, Andrew J. ;
Bach, Udo ;
Winkler, David A. ;
Russo, Salvy P. .
NPJ COMPUTATIONAL MATERIALS, 2020, 6 (01)