Transfer learning for chemically accurate interatomic neural network potentials

被引:21
作者
Zaverkin, Viktor [1 ]
Holzmueller, David [2 ]
Bonfirraro, Luca [1 ]
Kaestner, Johannes [1 ]
机构
[1] Univ Stuttgart, Inst Theoret Chem, Fac Chem, Stuttgart, Germany
[2] Univ Stuttgart, Inst Stochast & Applicat, Fac Math & Phys, Stuttgart, Germany
关键词
MACHINE; SYSTEMS; MODELS;
D O I
10.1039/d2cp05793j
中图分类号
O64 [物理化学(理论化学)、化学物理学];
学科分类号
070304 ; 081704 ;
摘要
Developing machine learning-based interatomic potentials from ab initio electronic structure methods remains a challenging task for computational chemistry and materials science. This work studies the capability of transfer learning, in particular discriminative fine-tuning, for efficiently generating chemically accurate interatomic neural network potentials on organic molecules from the MD17 and ANI data sets. We show that pre-training the network parameters on data obtained from density functional calculations considerably improves the sample efficiency of models trained on more accurate ab initio data. Additionally, we show that fine-tuning with energy labels alone can suffice to obtain accurate atomic forces and run large-scale atomistic simulations, provided a well-designed fine-tuning data set. We also investigate possible limitations of transfer learning, especially regarding the design and size of the pre-training and fine-tuning data sets. Finally, we provide GM-NN potentials pre-trained and fine-tuned on the ANI-1x and ANI-1ccx data sets, which can easily be fine-tuned on and applied to organic molecules.
引用
收藏
页码:5383 / 5396
页数:14
相关论文
共 69 条
  • [1] Abadi M., 2015, TensorFlow: LargeScale Machine Learning on Heterogeneous Systems." Software
  • [2] Coupled-cluster theory in quantum chemistry
    Bartlett, Rodney J.
    Musial, Monika
    [J]. REVIEWS OF MODERN PHYSICS, 2007, 79 (01) : 291 - 352
  • [3] Multifidelity Information Fusion with Machine Learning: A Case Study of Dopant Formation Energies in Hafnia
    Batra, Rohit
    Pilania, Ghanshyam
    Uberuaga, Blas P.
    Ramprasad, Rampi
    [J]. ACS APPLIED MATERIALS & INTERFACES, 2019, 11 (28) : 24906 - 24918
  • [4] Generalized neural-network representation of high-dimensional potential-energy surfaces
    Behler, Joerg
    Parrinello, Michele
    [J]. PHYSICAL REVIEW LETTERS, 2007, 98 (14)
  • [5] Brown T., 2020, P ADV NEUR INF PROC, V33, P1877
  • [6] Chen MS, 2022, Arxiv, DOI arXiv:2211.16619
  • [7] Chen Ting, 2019, 25 AMERICAS C INFORM
  • [8] Towards exact molecular dynamics simulations with machine-learned force fields
    Chmiela, Stefan
    Sauceda, Huziel E.
    Mueller, Klaus-Robert
    Tkatchenko, Alexandre
    [J]. NATURE COMMUNICATIONS, 2018, 9
  • [9] Machine learning of accurate energy-conserving molecular force fields
    Chmiela, Stefan
    Tkatchenko, Alexandre
    Sauceda, Huziel E.
    Poltavsky, Igor
    Schuett, Kristof T.
    Mueller, Klaus-Robert
    [J]. SCIENCE ADVANCES, 2017, 3 (05):
  • [10] On the role of gradients for machine learning of molecular energies and forces
    Christensen, Anders S.
    Von Lilienfeld, O. Anatole
    [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2020, 1 (04):