Learnable activation functions in physics-informed neural networks for solving partial differential equations

被引:0
作者
Farea, Afrah [1 ]
Celebi, Mustafa Serdar [1 ]
机构
[1] Istanbul Tech Univ, Dept Computat Sci & Engn, TR-34469 Istanbul, Turkiye
关键词
Partial differential equations; Physics informed neural networks; Multilayer perceptrons; Kolmogorov-Arnold networks; Learnable activation function;
D O I
10.1016/j.cpc.2025.109753
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Physics-Informed Neural Networks (PINNs) have emerged as a promising approach for solving Partial Differential Equations (PDEs). However, they face challenges related to spectral bias (the tendency to learn low-frequency components while struggling with high-frequency features) and unstable convergence dynamics (mainly stemming from the multi-objective nature of the PINN loss function). These limitations impact their accuracy for solving problems involving rapid oscillations, sharp gradients, and complex boundary behaviors. We systematically investigate learnable activation functions as a solution to these challenges, comparing Multilayer Perceptrons (MLPs) using fixed and learnable activation functions against Kolmogorov-Arnold Networks (KANs) that employ learnable basis functions. Our evaluation spans diverse PDE types, including linear and non-linear wave problems, mixed-physics systems, and fluid dynamics. Using empirical Neural Tangent Kernel (NTK) analysis and Hessian eigenvalue decomposition, we assess spectral bias and convergence stability of the models. Our results reveal a trade-off between expressivity and training convergence stability. While learnable activation functions work well in simpler architectures, they encounter scalability issues in complex networks due to the higher functional dimensionality. Counterintuitively, we find that low spectral bias alone does not guarantee better accuracy, as functions with broader NTK eigenvalue spectra may exhibit convergence instability. We demonstrate that activation function selection remains inherently problem-specific, with different bases showing distinct advantages for particular PDE characteristics. We believe these insights will help in the design of more robust neural PDE solvers.
引用
收藏
页数:18
相关论文
共 79 条
[1]  
Alain G., 2019, INT C LEARN REPR
[2]  
Heydari AA, 2019, Arxiv, DOI arXiv:1912.12355
[3]   Residual-based attention in physics-informed neural networks [J].
Anagnostopoulos, Sokratis J. ;
Toscano, Juan Diego ;
Stergiopulos, Nikolaos ;
Karniadakis, George Em .
COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2024, 421
[4]   A survey on modern trainable activation functions [J].
Apicella, Andrea ;
Donnarumma, Francesco ;
Isgro, Francesco ;
Prevete, Roberto .
NEURAL NETWORKS, 2021, 138 :14-32
[5]   Uncovering near-wall blood flow from sparse data with physics-informed neural networks [J].
Arzani, Amirhossein ;
Wang, Jian-Xun ;
D'Souza, Roshan M. .
PHYSICS OF FLUIDS, 2021, 33 (07)
[6]   Finite element interpolated neural networks for solving forward and inverse problems [J].
Badia, Santiago ;
Li, Wei ;
Martin, Alberto F. .
COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2024, 418
[7]   Physics-informed radial basis network (PIRBN): A local approximating neural network for solving nonlinear partial differential equations [J].
Bai, Jinshuai ;
Liu, Gui-Rong ;
Gupta, Ashish ;
Alzubaidi, Laith ;
Feng, Xi-Qiao ;
Gu, YuanTong .
COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2023, 415
[8]   Kronecker Neural Networks Overcome Spectral Bias for PINN-Based Wavefield Computation [J].
bin Waheed, Umair .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
[9]   Multi-Objective Loss Balancing for Physics-Informed Deep Learning [J].
Bischof, Rafael ;
Kraus, Michael A. .
COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2025, 439
[10]  
Bozorgasl Z, 2024, Arxiv, DOI [arXiv:2405.12832, DOI 10.48550/ARXIV.2405.12832]