Compositional design of multicomponent alloys using reinforcement learning

被引:6
|
作者
Xian, Yuehui [1 ]
Dang, Pengfei [1 ]
Tian, Yuan [1 ]
Jiang, Xue [2 ]
Zhou, Yumei [1 ]
Ding, Xiangdong [1 ]
Sun, Jun [1 ]
Lookman, Turab [1 ,2 ,3 ]
Xue, Dezhen [1 ]
机构
[1] Xi An Jiao Tong Univ, State Key Lab Mech Behav Mat, Xian 710049, Peoples R China
[2] Univ Sci & Technol Beijing, Beijing Adv Innovat Ctr Mat Genome Engn, Beijing 100083, Peoples R China
[3] AiMat Res LLC, Santa Fe, NM 87501 USA
基金
中国国家自然科学基金;
关键词
Compositional design; Reinforcement learning; Multicomponent alloys; Transformational enthalpy; Phase change materials; PHASE-CHANGE MATERIALS; HIGH ENTROPY ALLOYS; TEMPERATURES; STORAGE;
D O I
10.1016/j.actamat.2024.120017
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The design of alloys has typically involved adaptive experimental synthesis and characterization guided by machine learning models fitted to available data. A bottleneck for sequential design, be it for self-driven or manual synthesis, by Bayesian Global Optimization (BGO) for example, is that the search space becomes intractable as the number of alloy elements and its compositions exceed a threshold. Here we investigate how reinforcement learning (RL) performs in the compositional design of alloys within a closed loop with manual synthesis and characterization. We demonstrate this strategy by designing a phase change multicomponent alloy (Ti 27.2 Ni 47 Hf 13.8 Zr 12 ) with the highest transformation enthalpy (Delta H) Delta H)-37.1 J/g (-39.0 J/g with further calibration) within the TiNi-based family of alloys from a space of over 2 x 108 8 candidates, although the initial training is only on a compact dataset of 112 alloys. We show how the training efficiency is increased by employing acquisition functions containing uncertainties, such as expected improvement (EI), as the reward itself. Existing alloy data is often limited, however, if the agent is pretrained on experimental results prior to the training process, it can access regions of higher reward values more frequently. In addition, the experimental feedback enables the agent to gradually explore new regions with higher rewards, compositionally different from the initial dataset. Our approach directly applies to processing conditions where the actions would be performed in a given order. We also compare RL performance to BGO and the genetic algorithm on several test functions to gain insight on their relative strengths in materials design.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Reinforcement Learning inspired Deep Learned Compositional Model for Decision Making in Tracking
    Chakraborty, Anit
    Dutta, Sayandip
    Bhattacharyya, Siddhartha
    Platos, Jan
    Snasel, Vaclav
    2018 FOURTH IEEE INTERNATIONAL CONFERENCE ON RESEARCH IN COMPUTATIONAL INTELLIGENCE AND COMMUNICATION NETWORKS (ICRCICN), 2018, : 158 - 163
  • [42] Monopoly Using Reinforcement Learning
    Arun, Edupuganti
    Rajesh, Harikrishna
    Chakrabarti, Debarka
    Cherala, Harikiran
    George, Koshy
    PROCEEDINGS OF THE 2019 IEEE REGION 10 CONFERENCE (TENCON 2019): TECHNOLOGY, KNOWLEDGE, AND SOCIETY, 2019, : 864 - 868
  • [43] Dynamic retail market tariff design for an electricity aggregator using reinforcement learning
    Naseri, Nastaran
    Talari, Saber
    Ketter, Wolfgang
    Collins, John
    ELECTRIC POWER SYSTEMS RESEARCH, 2022, 212
  • [44] A novel design of hidden web crawler using reinforcement learning based agents
    Akilandeswari, J.
    Gopalan, N. P.
    ADVANCED PARALLEL PROCESSING TECHNOLOGIES, PROCEEDINGS, 2007, 4847 : 433 - +
  • [45] Automatic Facility Layout Design Using Reinforcement Learning and a Analytic Hierarchy Process
    Ikeda H.
    Nakagawa H.
    Akagi H.
    Sekimoto F.
    Tsuchiya T.
    Journal of Japan Industrial Management Association, 2023, 74 (03) : 142 - 152
  • [46] FastTuner: Transferable Physical Design Parameter Optimization using Fast Reinforcement Learning
    Hsiao, Hao-Hsiang
    Lu, Yi-Chen
    Vanna-Iampikul, Pruek
    Lim, Sung Kyu
    PROCEEDINGS OF THE 2024 INTERNATIONAL SYMPOSIUM ON PHYSICAL DESIGN, ISPD 2024, 2024, : 93 - 101
  • [47] Joint Cache Placement and Delivery Design using Reinforcement Learning for Cellular Networks
    Amidzadeh, Mohsen
    Al-Tous, Hanan
    Tirkkonen, Olav
    Zhang, Junshan
    2021 IEEE 93RD VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-SPRING), 2021,
  • [48] Hierarchical Multiresolution Design of Bioinspired Structural Composites Using Progressive Reinforcement Learning
    Yu, Chi-Hua
    Tseng, Bor-Yann
    Yang, Zhenze
    Tung, Cheng-Che
    Zhao, Elena
    Ren, Zhi-Fan
    Yu, Sheng-Sheng
    Chen, Po-Yu
    Chen, Chuin-Shan
    Buehler, Markus J.
    ADVANCED THEORY AND SIMULATIONS, 2022, 5 (11)
  • [49] Adaptive Design Parameter Determination for Control Barrier Functions using Reinforcement Learning
    Memis, Sezer
    Demir, Esra
    Senel, Serkan
    Demir, Mustafa
    Koyuncu, Emre
    IFAC PAPERSONLINE, 2024, 58 (30): : 186 - 191
  • [50] Using multicomponent recycled electronic waste alloys to produce high entropy alloys
    Torralba, Jose M.
    Iriarte, Diego
    Tourret, Damien
    Meza, Alberto
    INTERMETALLICS, 2024, 164