Enhancing reinforcement learning for de novo molecular design applying self-attention mechanisms

被引:1
|
作者
Pereira, Tiago O. [1 ]
Abbasi, Maryam [2 ]
Arrais, Joel P. [3 ]
机构
[1] Univ Coimbra, Dept Informat Engn, Informat Engn, Coimbra, Portugal
[2] Univ Coimbra, Ctr Informat & Syst, Coimbra, Portugal
[3] Univ Coimbra, Dept Informat Engn, Coimbra, Portugal
关键词
drug design; smiles; deep learning; transformer; reinforcement learning; GENERATION; USP7; CANCER; TRANSFORMER; LIBRARIES;
D O I
10.1093/bib/bbad368
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
The drug discovery process can be significantly improved by applying deep reinforcement learning (RL) methods that learn to generate compounds with desired pharmacological properties. Nevertheless, RL-based methods typically condense the evaluation of sampled compounds into a single scalar value, making it difficult for the generative agent to learn the optimal policy. This work combines self-attention mechanisms and RL to generate promising molecules. The idea is to evaluate the relative significance of each atom and functional group in their interaction with the target, and to utilize this information for optimizing the Generator. Therefore, the framework for de novo drug design is composed of a Generator that samples new compounds combined with a Transformer-encoder and a biological affinity Predictor that evaluate the generated structures. Moreover, it takes the advantage of the knowledge encapsulated in the Transformer's attention weights to evaluate each token individually. We compared the performance of two output prediction strategies for the Transformer: standard and masked language model (MLM). The results show that the MLM Transformer is more effective in optimizing the Generator compared with the state-of-the-art works. Additionally, the evaluation models identified the most important regions of each molecule for the biological interaction with the target. As a case study, we generated synthesizable hit compounds that can be putative inhibitors of the enzyme ubiquitin-specific protein 7 (USP7).
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Reinforcement Learning with Self-Attention Networks for Cryptocurrency Trading
    Betancourt, Carlos
    Chen, Wen-Hui
    APPLIED SCIENCES-BASEL, 2021, 11 (16):
  • [2] Molecular de-novo design through deep reinforcement learning
    Olivecrona, Marcus
    Blaschke, Thomas
    Engkvist, Ola
    Chen, Hongming
    JOURNAL OF CHEMINFORMATICS, 2017, 9
  • [3] De novo molecular design using deep reinforcement learning methods
    Chen, Hongming
    Olivercrona, Markus
    Blaschke, Thomas
    Engkvist, Ola
    Kogej, Thierry
    Tyrchan, Christian
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2018, 256
  • [4] Molecular de-novo design through deep reinforcement learning
    Marcus Olivecrona
    Thomas Blaschke
    Ola Engkvist
    Hongming Chen
    Journal of Cheminformatics, 9
  • [5] A Self-attention Agent of Reinforcement Learning in Continuous Integration Testing
    Liu, Bangfu
    Li, Zheng
    Zhao, Ruilian
    Shang, Ying
    2023 IEEE 47TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC, 2023, : 886 - 891
  • [6] Adversarial Inverse Reinforcement Learning With Self-Attention Dynamics Model
    Sun, Jiankai
    Yu, Lantao
    Dong, Pinqian
    Lu, Bo
    Zhou, Bolei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 1880 - 1886
  • [7] Memory-assisted reinforcement learning for diverse molecular de novo design
    Blaschke, Thomas
    Engkvist, Ola
    Bajorath, Juergen
    Chen, Hongming
    JOURNAL OF CHEMINFORMATICS, 2020, 12 (01)
  • [8] Memory-assisted reinforcement learning for diverse molecular de novo design
    Thomas Blaschke
    Ola Engkvist
    Jürgen Bajorath
    Hongming Chen
    Journal of Cheminformatics, 12
  • [9] Applying Self-attention for Stance Classification
    Bugueno, Margarita
    Mendoza, Marcelo
    PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS (CIARP 2019), 2019, 11896 : 51 - 61
  • [10] Utilizing reinforcement learning for de novo drug design
    Gummesson Svensson, Hampus
    Tyrchan, Christian
    Engkvist, Ola
    Haghir Chehreghani, Morteza
    MACHINE LEARNING, 2024, 113 (07) : 4811 - 4843