Positional embeddings and zero-shot learning using BERT for molecular-property prediction

被引:0
|
作者
Mswahili, Medard Edmund [1 ]
Hwang, Junha [1 ]
Rajapakse, Jagath C. [2 ]
Jo, Kyuri [1 ]
Jeong, Young-Seob [1 ]
机构
[1] Chungbuk Natl Univ, Dept Comp Engn, Cheongju 28644, South Korea
[2] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
来源
JOURNAL OF CHEMINFORMATICS | 2025年 / 17卷 / 01期
关键词
Transformers; BERT; Positional embedding/encoding; Zero-shot learning; Molecular-property prediction; SMILES; DeepSMILES; GRAPH NEURAL-NETWORK; DRUG DISCOVERY; MODELS; ALGORITHM; DATABASE;
D O I
10.1186/s13321-025-00959-9
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Recently, advancements in cheminformatics such as representation learning for chemical structures, deep learning (DL) for property prediction, data-driven discovery, and optimization of chemical data handling, have led to increased demands for handling chemical simplified molecular input line entry system (SMILES) data, particularly in text analysis tasks. These advancements have driven the need to optimize components like positional encoding and positional embeddings (PEs) in transformer model to better capture the sequential and contextual information embedded in molecular representations. SMILES data represent complex relationships among atoms or elements, rendering them critical for various learning tasks within the field of cheminformatics. This study addresses the critical challenge of encoding complex relationships among atoms in SMILES strings to explore various PEs within the transformer-based framework to increase the accuracy and generalization of molecular property predictions. The success of transformer-based models, such as the bidirectional encoder representations from transformer (BERT) models, in natural language processing tasks has sparked growing interest from the domain of cheminformatics. However, the performance of these models during pretraining and fine-tuning is significantly influenced by positional information such as PEs, which help in understanding the intricate relationships within sequences. Integrating position information within transformer architectures has emerged as a promising approach. This encoding mechanism provides essential supervision for modeling dependencies among elements situated at different positions within a given sequence. In this study, we first conduct pretraining experiments using various PEs to explore diverse methodologies for incorporating positional information into the BERT model for chemical text analysis using SMILES strings. Next, for each PE, we fine-tune the best-performing BERT (masked language modeling) model on downstream tasks for molecular-property prediction. Here, we use two molecular representations, SMILES and DeepSMILES, to comprehensively assess the potential and limitations of the PEs in zero-shot learning analysis, demonstrating the model's proficiency in predicting properties of unseen molecular representations in the context of newly proposed and existing datasets.Scientific contributionThis study explores the unexplored potential of PEs using BERT model for molecular property prediction. The study involved pretraining and fine-tuning the BERT model on various datasets related to COVID-19, bioassay data, and other molecular and biological properties using SMILES and DeepSMILES representations. The study details the pretraining architecture, fine-tuning datasets, and the performance of the BERT model with different PEs. It also explores zero-shot learning analysis and the model's performance on various classification and regression tasks. In this study, newly proposed datasets from different domains were introduced during fine-tuning in addition to the existing and commonly used datasets. The study highlights the robustness of the BERT model in predicting chemical properties and its potential applications in cheminformatics and bioinformatics.
引用
收藏
页数:22
相关论文
共 50 条
  • [1] LEARNING VISUALLY CONSISTENT LABEL EMBEDDINGS FOR ZERO-SHOT LEARNING
    Demirel, Berkan
    Cinbis, Ramazan Gokberk
    Ikizler-Cinbis, Nazli
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3656 - 3660
  • [2] Zero-Shot Audio Classification using Image Embeddings
    Dogan, Duygu
    Xie, Huang
    Heittola, Toni
    Virtanen, Tuomas
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1 - 5
  • [3] Zero-Shot Visual Emotion Recognition by Exploiting BERT
    Kang, Hyunwook
    Hazarika, Devamanyu
    Kim, Dongho
    Kim, Jihie
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, 2023, 543 : 485 - 494
  • [4] mcBERT: Momentum Contrastive Learning with BERT for Zero-Shot Slot Filling
    Heo, Seong-Hwan
    Lee, WonKee
    Lee, Jong-Hyeok
    INTERSPEECH 2022, 2022, : 1243 - 1247
  • [5] Zero-Shot Learning for IMU-Based Activity Recognition Using Video Embeddings
    Tong, Catherine
    Ge, Jinchen
    Lane, Nicholas D.
    PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2021, 5 (04):
  • [6] Zero-shot learning with self-supervision by shuffling semantic embeddings
    Kim, Hoseong
    Lee, Jewook
    Byun, Hyeran
    NEUROCOMPUTING, 2021, 437 : 1 - 8
  • [7] Visual Context Embeddings for Zero-Shot Recognition
    Cho, Gunhee
    Choi, Yong Suk
    37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 1039 - 1047
  • [8] Joint attribute chain prediction for zero-shot learning
    Qiao, Lingfeng
    Tuo, Hongya
    Wang, Jiexin
    Wang, Chao
    Jing, Zhongliang
    IET COMPUTER VISION, 2018, 12 (06) : 873 - 881
  • [9] Scalable Zero-Shot Learning via Binary Visual-Semantic Embeddings
    Shen, Fumin
    Zhou, Xiang
    Yu, Jun
    Yang, Yang
    Liu, Li
    Shen, Heng Tao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (07) : 3662 - 3674
  • [10] Few-shot learning with transformers via graph embeddings for molecular property prediction
    Torres, Luis H. M.
    Ribeiro, Bernardete
    Arrais, Joel P.
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 225