Neural Network with Optimal Neuron Activation Functions Based on Additive Gaussian Process Regression

被引:18
作者
Manzhos, Sergei [1 ]
Ihara, Manabu [1 ]
机构
[1] Tokyo Inst Technol, Sch Mat & Chem Technol, Meguro Ku, Tokyo 1528552, Japan
关键词
UNIVERSAL APPROXIMATION; NUMERICAL IMPLEMENTATION; SUPERPOSITION THEOREM; RS-HDMR; ENERGY; MACHINE; REPRESENTATION; SIMULATIONS; VARIABLES; BOUNDS;
D O I
10.1021/acs.jpca.3c02949
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Feed-forward neural networks (NNs) are a staple machine learning method widely used in many areas of science and technology, including physical chemistry, computational chemistry, and materials informatics. While even a single-hidden-layer NN is a universal approximator, its expressive power is limited by the use of simple neuron activation functions (such as sigmoid functions) that are typically the same for all neurons. More flexible neuron activation functions would allow the use of fewer neurons and layers and thereby save computational cost and improve expressive power. We show that additive Gaussian process regression (GPR) can be used to construct optimal neuron activation functions that are individual to each neuron. An approach is also introduced that avoids nonlinear fitting of neural network parameters by defining them with rules. The resulting method combines the advantage of robustness of a linear regression with the higher expressive power of an NN. We demonstrate the approach by fitting the potential energy surfaces of the water molecule and formaldehyde. Without requiring any nonlinear optimization, the additive-GPR-based approach outperforms a conventional NN in the high-accuracy regime, where a conventional NN suffers more from overfitting.
引用
收藏
页码:7823 / 7835
页数:13
相关论文
共 101 条
[1]   Efficient implementation of high dimensional model representations [J].
Alis, ÖF ;
Rabitz, H .
JOURNAL OF MATHEMATICAL CHEMISTRY, 2001, 29 (02) :127-142
[2]   Convolutional Neural Network of Atomic Surface Structures To Predict Binding Energies for High-Throughput Screening of Catalysts [J].
Back, Seoin ;
Yoon, Junwoong ;
Tian, Nianhan ;
Zhong, Wen ;
Tran, Kevin ;
Ulissi, Zachary W. .
JOURNAL OF PHYSICAL CHEMISTRY LETTERS, 2019, 10 (15) :4401-4408
[3]   UNIVERSAL APPROXIMATION BOUNDS FOR SUPERPOSITIONS OF A SIGMOIDAL FUNCTION [J].
BARRON, AR .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1993, 39 (03) :930-945
[4]   The multiconfiguration time-dependent Hartree (MCTDH) method:: a highly efficient algorithm for propagating wavepackets [J].
Beck, MH ;
Jäckle, A ;
Worth, GA ;
Meyer, HD .
PHYSICS REPORTS-REVIEW SECTION OF PHYSICS LETTERS, 2000, 324 (01) :1-105
[5]   Representing potential energy surfaces by high-dimensional neural network potentials [J].
Behler, J. .
JOURNAL OF PHYSICS-CONDENSED MATTER, 2014, 26 (18)
[6]   Generalized neural-network representation of high-dimensional potential-energy surfaces [J].
Behler, Joerg ;
Parrinello, Michele .
PHYSICAL REVIEW LETTERS, 2007, 98 (14)
[7]   First Principles Neural Network Potentials for Reactive Simulations of Large Molecular and Condensed Systems [J].
Behler, Joerg .
ANGEWANDTE CHEMIE-INTERNATIONAL EDITION, 2017, 56 (42) :12828-12840
[8]   Perspective: Machine learning potentials for atomistic simulations [J].
Behler, Joerg .
JOURNAL OF CHEMICAL PHYSICS, 2016, 145 (17)
[9]   Constructing high-dimensional neural network potentials: A tutorial review [J].
Behler, Joerg .
INTERNATIONAL JOURNAL OF QUANTUM CHEMISTRY, 2015, 115 (16) :1032-1050
[10]   Atomistic simulations of thermal conductivity in GeTe nanowires [J].
Bosoni, E. ;
Campi, D. ;
Donadio, D. ;
Sosso, G. C. ;
Behler, J. ;
Bernasconi, M. .
JOURNAL OF PHYSICS D-APPLIED PHYSICS, 2020, 53 (05)