From Static to Dynamic: Knowledge Metabolism for Large Language Models

被引:0
作者
Du, Mingzhe [1 ,2 ]
Luu, Anh Tuan [1 ]
Ji, Bin [2 ]
Ng, See-Kiong [2 ]
机构
[1] Nanyang Technol Univ, Singapore, Singapore
[2] Natl Univ Singapore, Singapore, Singapore
来源
THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21 | 2024年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The immense parameter space of Large Language Models (LLMs) endows them with superior knowledge retention capabilities, allowing them to excel in a variety of natural language processing tasks. However, it also instigates difficulties in consistently tuning LLMs to incorporate the most recent knowledge, which may further lead LLMs to produce inaccurate and fabricated content. To alleviate this issue, we propose a knowledge metabolism framework for LLMs, which proactively sustains the credibility of knowledge through an auxiliary memory component and directly delivers pertinent knowledge for LLM inference, thereby suppressing hallucinations caused by obsolete internal knowledge during the LLM inference process. Benchmark experiments demonstrate DynaMind's effectiveness in overcoming this challenge. The code and demo of DynaMind are available at: https://github.com/Elfsong/DynaMind.
引用
收藏
页码:23784 / 23786
页数:3
相关论文
共 50 条
[41]   SKILL: Structured Knowledge Infusion for Large Language Models [J].
Moiseev, Fedor ;
Dong, Zhe ;
Alfonseca, Enrique ;
Jaggi, Martin .
NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, :1581-1588
[42]   Knowledge Graph Treatments for Hallucinating Large Language Models [J].
Collarana, Diego ;
Busch, Moritz ;
Lange, Christoph .
ERCIM NEWS, 2024, (136) :35-36
[43]   LawBench: Benchmarking Legal Knowledge of Large Language Models [J].
Fei, Zhiwei ;
Shen, Xiaoyu ;
Zhu, Dawei ;
Zhou, Fengzhe ;
Han, Zhuo ;
Huang, Alan ;
Zhang, Songyang ;
Chen, Kai ;
Yin, Zhixin ;
Shen, Zongwen ;
Ge, Jidong ;
Ng, Vincent .
2024 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2024, 2024, :7933-7962
[44]   A Survey on Symbolic Knowledge Distillation of Large Language Models [J].
Acharya, Kamal ;
Velasquez, Alvaro ;
Song, Houbing Herbert .
IEEE Transactions on Artificial Intelligence, 2024, 5 (12) :5928-5948
[45]   Detoxifying Large Language Models via Knowledge Editing [J].
Wang, Mengru ;
Zhang, Ningyu ;
Xu, Ziwen ;
Xi, Zekun ;
Deng, Shumin ;
Yao, Yunzhi ;
Zhang, Qishen ;
Yang, Linyi ;
Wang, Jindong ;
Chen, Huajun .
PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, :3093-3118
[46]   Knowledge-driven Scientific Large Language Models [J].
Zhang, Qiang .
THIRTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, AAAI-25, VOL 39 NO 27, 2025, :28736-28736
[47]   Unifying Large Language Models and Knowledge Graphs: A Roadmap [J].
Pan, Shirui ;
Luo, Linhao ;
Wang, Yufei ;
Chen, Chen ;
Wang, Jiapu ;
Wu, Xindong .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (07) :3580-3599
[48]   Enhanced Story Comprehension for Large Language Models through Dynamic Document-Based Knowledge Graphs [J].
Andrus, Berkeley R. ;
Nasiri, Yeganeh ;
Cui, Shilong ;
Cullen, Benjamin ;
Fulda, Nancy .
THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, :10436-10444
[49]   Dynamic Voting for Efficient Reasoning in Large Language Models [J].
Xue, Mingfeng ;
Liu, Dayiheng ;
Lei, Wenqiang ;
Ren, Xingzhang ;
Yang, Baosong ;
Xie, Jun ;
Zhang, Yidan ;
Peng, Dezhong ;
Lv, Jiancheng .
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, :3085-3104
[50]   Extracting individual trajectories from text by fusing large language models with diverse knowledge [J].
Liu, Le ;
Pei, Tao ;
Fang, Zidong ;
Yan, Xiaorui ;
Zheng, Chenglong ;
Wang, Xi ;
Song, Ci ;
Luan, Wenfei ;
Chen, Jie .
INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2025, 141