A multi-fidelity information-fusion approach to machine learn and predict polymer bandgap

被引:72
作者
Patra, Abhirup [1 ]
Batra, Rohit [1 ]
Chandrasekaran, Anand [1 ]
Kim, Chiho [1 ]
Tran Doan Huan [1 ]
Ramprasad, Rampi [1 ]
机构
[1] Georgia Inst Technol, Sch Mat Sci & Engn, 771 Ferst Dr NW, Atlanta, GA 30332 USA
关键词
Density functional theory; Machine learning; Multi-fidelity learning; Polymers; Bandgap; APPROXIMATION; POTENTIALS;
D O I
10.1016/j.commatsci.2019.109286
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The fidelity of data is of paramount importance in the construction of reliable and accurate machine learning (ML) models. Low-fidelity data, although noisy, can usually be obtained for a large number of material samples. High-fidelity data, on the other hand, is time-consuming and oftentimes, only available for a limited number of target samples. While the former can provide useful information to help generalize the ML models over large materials space, the latter is useful to build more accurate surrogate models. Information fusion schemes that utilize the data available at multiple levels of fidelity can outperform traditional single fidelity based ML methods, such as Gaussian process regression. In this work, a variant of the multi-fidelity information fusion scheme, namely multi-fidelity co-kriging, is used to build powerful prediction models of polymer bandgaps. To benchmark this strategy, we utilize a bandgap dataset of 382 polymers, obtained at two levels of fidelity: using the Perdew-Burke-Ernzerhof (PBE) exchange-correlational functional ("low-fidelity") and the Heyd-Scuseria-Ernzerhof (HSE06) functional("high-fidelity") of density functional theory. The multi-fidelity model, trained on both PBE and HSE06 data, outperforms a single-fidelity Gaussian process regression model trained on just HSE06 band-gaps in a number of scenarios and is also able to generalize better to a more diverse chemical space.
引用
收藏
页数:5
相关论文
共 31 条
[1]  
[Anonymous], 2017, NPJ COMPUT MATER
[2]  
[Anonymous], 2018, MATER TODAY, DOI DOI 10.1016/j.mattod.2017.11.021
[3]  
[Anonymous], ACS APPL MAT INTERFA
[4]   Gaussian approximation potentials: A brief tutorial introduction [J].
Bartok, Albert P. ;
Csanyi, Gabor .
INTERNATIONAL JOURNAL OF QUANTUM CHEMISTRY, 2015, 115 (16) :1051-1057
[5]  
Bishop C. M., 2006, Pattern Recognition and Machine Learning
[6]   Machine learning for molecular and materials science [J].
Butler, Keith T. ;
Davies, Daniel W. ;
Cartwright, Hugh ;
Isayev, Olexandr ;
Walsh, Aron .
NATURE, 2018, 559 (7715) :547-555
[7]  
Carrasquilla J, 2017, NAT PHYS, V13, P431, DOI [10.1038/NPHYS4035, 10.1038/nphys4035]
[8]   Solving the electronic structure problem with machine learning [J].
Chandrasekaran, Anand ;
Kamal, Deepak ;
Batra, Rohit ;
Kim, Chiho ;
Chen, Lihua ;
Ramprasad, Rampi .
NPJ COMPUTATIONAL MATERIALS, 2019, 5 (1)
[9]   Multi-fidelity optimization via surrogate modelling [J].
Forrester, Alexander I. J. ;
Sobester, Andras ;
Keane, Andy J. .
PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2007, 463 (2088) :3251-3269
[10]   Machine Learning Predictions of Molecular Properties: Accurate Many-Body Potentials and Nonlocality in Chemical Space [J].
Hansen, Katja ;
Biegler, Franziska ;
Ramakrishnan, Raghunathan ;
Pronobis, Wiktor ;
von Lilienfeld, O. Anatole ;
Mueller, Klaus-Robert ;
Tkatchenko, Alexandre .
JOURNAL OF PHYSICAL CHEMISTRY LETTERS, 2015, 6 (12) :2326-2331