A novel approach to unlocking the synergy of large language models and chemical knowledge in biomedical signal applications

被引:2
作者
Yin, Zilong [1 ]
Wang, Haoyu [1 ]
Chen, Bin [1 ,4 ]
Sun, Hangling [2 ]
Li, Anji [3 ]
Zhou, Chenyu [5 ,6 ]
机构
[1] Univ Shanghai Sci & Technol, Shanghai, Peoples R China
[2] Hengtu Imalligent Technol Shanghai Co Ltd, Shanghai, Peoples R China
[3] Abbott Labs Shanghai Co Ltd, Shanghai, Peoples R China
[4] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[5] Xinjiang Univ, Urumqi, Peoples R China
[6] Tsinghua Univ, Beijing, Peoples R China
关键词
Biomedical signal processing; Supervised chemical knowledge; Large language models; Molecular property prediction; PREDICTION; EXPLORATION; ENTITIES; DATABASE;
D O I
10.1016/j.bspc.2024.107388
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
This work explores the potential of using the pre-trained large language model Llama2 to address challenges in biomedical signal processing and control (BSPC), particularly in predicting the electronic and functional properties of organic molecules, an area of growing importance infields such as drug discovery and materials science. Current approaches in BSPC often rely on specialized graph neural network models, which can be limited in their ability to capture the complex relationships inherent in molecular structures. To address this, we demonstrate that a fine-tuned Llama2 model can accurately predict the highest occupied molecular orbital (HOMO) and lowest unoccupied molecular orbital (LUMO) energies of organic semiconductor molecules, with performance comparable to state-of-the-art specialized models. To further enhance the model's robustness and generalization, we propose several key innovations, including optimized simplified molecular input line entry system (SMILES) tokenization, incorporation of chemical knowledge as auxiliary supervised tasks, and a low-rank adaptation (LORA) based fine-tuning strategy. These techniques enable the language model to simultaneously learn SMILES prediction and acquire relevant chemical knowledge, while also improving its handling of incomplete structural information and ability to generalize to "unseen" molecular classes. The work also discusses the limitations of using large language models for molecular property prediction, such as the lack of interpretability and the need for improved handling of non-standard SMILES representations, highlighting the potential of this approach in BSPC while identifying areas for further improvement.
引用
收藏
页数:16
相关论文
共 50 条
[31]   To take a different approach: Can large language models provide knowledge related to respiratory aspiration? [J].
Niu, Yirou ;
Fu, Shuojin ;
Xuan, Zehui ;
Ren, Zhifang ;
Jin, Shuai ;
Wang, Yanling ;
Xiao, Qian .
DIGITAL HEALTH, 2025, 11
[32]   Adapting Large Language Models to Biomedical Domain: A Survey of Techniques and Approaches [J].
Klila, Jaafer ;
Souihi, Sondes Bannour ;
Boujelben, Rahma ;
Semmar, Nasredine ;
Belguith, Lamia Hadrich .
ADVANCEMENTS IN MACHINE LEARNING AND NATURAL LANGUAGE PROCESSING: INNOVATIONS AND APPLICATIONS, LPKM'2024, 2025, 1303 :155-163
[33]   MedKA: A knowledge graph-augmented approach to improve factuality in medical Large Language Models [J].
Deng, Yiyan ;
Zhao, Shen ;
Miao, Yongming ;
Zhu, Junjie ;
Li, Jin .
JOURNAL OF BIOMEDICAL INFORMATICS, 2025, 168
[34]   ChatGPT, Bard, and Large Language Models for Biomedical Research: Opportunities and Pitfalls [J].
Thapa, Surendrabikram ;
Adhikari, Surabhi .
ANNALS OF BIOMEDICAL ENGINEERING, 2023, 51 (12) :2647-2651
[35]   Large Language Models in Biomedical and Health Informatics: A Review with Bibliometric Analysis [J].
Yu, Huizi ;
Fan, Lizhou ;
Li, Lingyao ;
Zhou, Jiayan ;
Ma, Zihui ;
Xian, Lu ;
Hua, Wenyue ;
He, Sijia ;
Jin, Mingyu ;
Zhang, Yongfeng ;
Gandhi, Ashvin ;
Ma, Xin .
JOURNAL OF HEALTHCARE INFORMATICS RESEARCH, 2024, 8 (04) :658-711
[36]   ChatGPT, Bard, and Large Language Models for Biomedical Research: Opportunities and Pitfalls [J].
Surendrabikram Thapa ;
Surabhi Adhikari .
Annals of Biomedical Engineering, 2023, 51 :2647-2651
[37]   Enhancing Relation Extraction from Biomedical Texts by Large Language Models [J].
Asada, Masaki ;
Fukuda, Ken .
ARTIFICIAL INTELLIGENCE IN HCI, PT III, AI-HCI 2024, 2024, 14736 :3-14
[38]   Rethinking chemical research in the age of large language models [J].
Macknight, Robert ;
Boiko, Daniil A. ;
Regio, Jose Emilio ;
Gallegos, Liliana C. ;
Neukomm, Theo A. ;
Gomes, Gabe .
NATURE COMPUTATIONAL SCIENCE, 2025,
[39]   Applications of Large Language Models and Multimodal Large Models in Autonomous Driving: A Comprehensive Review [J].
Li, Jing ;
Li, Jingyuan ;
Yang, Guo ;
Yang, Lie ;
Chi, Haozhuang ;
Yang, Lichao .
DRONES, 2025, 9 (04)
[40]   Large language models: unlocking new potential in patient education for thyroid eye disease [J].
Gao, Yuwan ;
Xu, Qi ;
Zhang, Ou ;
Wang, Hongliang ;
Wang, Yunlong ;
Wang, Jiale ;
Chen, Xiaohui .
ENDOCRINE, 2025,