De novo drug design based on Stack-RNN with multi-objective reward-weighted sum and reinforcement learning

被引:10
|
作者
Hu, Pengwei [1 ,2 ]
Zou, Jinping [1 ,2 ]
Yu, Jialin [1 ,2 ]
Shi, Shaoping [1 ,2 ]
机构
[1] Nanchang Univ, Sch Math & Comp Sci, Dept Math, Nanchang 330031, Peoples R China
[2] Nanchang Univ, Inst Math & Interdisciplinary Sci, Nanchang 330031, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; De novo drug; Multi-objective; Drug design; Molecular generation; NEURAL-NETWORKS; VALIDATION; RECEPTORS; MOLECULES;
D O I
10.1007/s00894-023-05523-6
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
ContextIn recent decades, drug development has become extremely important as different new diseases have emerged. However, drug discovery is a long and complex process with a very low success rate, and methods are needed to improve the efficiency of the process and reduce the possibility of failure. Among them, drug design from scratch has become a promising approach. Molecules are generated from scratch, reducing the reliance on trial and error and prefabricated molecular repositories, but the optimization of its molecular properties is still a challenging multi-objective optimization problem.MethodsIn this study, two stack-augmented recurrent neural networks were used to compose a generative model for generating drug-like molecules, and then reinforcement learning was used for optimization to generate molecules with desirable properties, such as binding affinity and the logarithm of the partition coefficient between octanol and water. In addition, a memory storage network was added to increase the internal diversity of the generated molecules. For multi-objective optimization, we proposed a new approach which utilized the magnitude of different attribute reward values to assign different weights to molecular optimization. The proposed model not only solves the problem that the properties of the generated molecules are extremely biased towards a certain attribute due to the possible conflict between the attributes, but also improves various properties of the generated molecules compared with the traditional weighted sum and alternating weighted sum, among which the molecular validity reaches 97.3%, the internal diversity is 0.8613, and the desirable molecules increases from 55.9 to 92%.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Virtual machine placement based on multi-objective reinforcement learning
    Qin, Yao
    Wang, Hua
    Yi, Shanwen
    Li, Xiaole
    Zhai, Linbo
    APPLIED INTELLIGENCE, 2020, 50 (08) : 2370 - 2383
  • [32] An Improved Multi-objective Optimization Algorithm Based on Reinforcement Learning
    Liu, Jun
    Zhou, Yi
    Qiu, Yimin
    Li, Zhongfeng
    ADVANCES IN SWARM INTELLIGENCE, ICSI 2022, PT I, 2022, : 501 - 513
  • [33] A multi-objective optimisation approach to the design of experiment in de novo assembly projects
    Nadalin, Francesca
    Vezzi, Francesco
    Policriti, Alberto
    2012 23RD INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS (DEXA), 2012, : 213 - 217
  • [34] Incremental reinforcement learning for multi-objective analog circuit design acceleration
    Abuelnasr, Ahmed
    Ragab, Ahmed
    Amer, Mostafa
    Gosselin, Benoit
    Savaria, Yvon
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 129
  • [35] Examining multi-objective deep reinforcement learning frameworks for molecular design
    Al-Jumaily, Aws
    Mukaidaisi, Muhetaer
    Vu, Andrew
    Tchagang, Alain
    Li, Yifeng
    BIOSYSTEMS, 2023, 232
  • [36] Predicting optimal value functions by interpolating reward functions in scalarized multi-objective reinforcement learning
    Kusari, Arpan
    How, Jonathan P.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 7484 - 7490
  • [37] Application Placement in Fog Environments using Multi-Objective Reinforcement Learning with Maximum Reward Formulation
    Eyckerman, Reinout
    Reiter, Phil
    Latre, Steven
    Marquez-Barja, Johann
    Hellinckx, Peter
    PROCEEDINGS OF THE IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM 2022, 2022,
  • [38] Reward-Balancing for Statistical Spoken Dialogue Systems using Multi-objective Reinforcement Learning
    Ultes, Stefan
    Budzianowski, Pawel
    Casanueva, Inigo
    Mrksic, Nikola
    Rojas-Barahona, Lina
    Su, Pei-Hao
    Wen, Tsung-Hsien
    Gasic, Milica
    Young, Steve
    18TH ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2017), 2017, : 65 - 70
  • [39] Deep reinforcement learning for multi-objective optimization in BIM-based green building design
    Pan, Yue
    Shen, Yuxuan
    Qin, Jianjun
    Zhang, Limao
    AUTOMATION IN CONSTRUCTION, 2024, 166
  • [40] Activity cliff-aware reinforcement learning for de novo drug design
    Xiuyuan Hu
    Guoqing Liu
    Yang Zhao
    Hao Zhang
    Journal of Cheminformatics, 17 (1)