Cross-Lingual Knowledge Editing in Large Language Models

被引:0
作者
Wang, Jiaan [1 ]
Liang, Yunlong [2 ]
Sun, Zengkui [3 ]
Cao, Yuxuan [4 ]
Xu, Jiarong [1 ]
Meng, Fandong [2 ]
机构
[1] Fudan Univ, Shanghai, Peoples R China
[2] Tencent Inc, Pattern Recognit Ctr, WeChat AI, Shenzhen, Peoples R China
[3] Beijing Jiaotong Univ, Beijing, Peoples R China
[4] Zhejiang Univ, Hangzhou, Peoples R China
来源
PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS | 2024年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge editing aims to change language models' performance on several special cases (i.e., editing scope) by infusing the corresponding expected knowledge into them. With the recent advancements in large language models (LLMs), knowledge editing has been shown as a promising technique to adapt LLMs to new knowledge without retraining from scratch. However, most of the previous studies neglect the multi-lingual nature of some main-stream LLMs (e.g., LLaMA, ChatGPT and GPT-4), and typically focus on monolingual scenarios, where LLMs are edited and evaluated in the same language. As a result, it is still unknown the effect of source language editing on a different target language. In this paper, we aim to figure out this cross-lingual effect in knowledge editing. Specifically, we first collect a largescale cross-lingual synthetic dataset by translating ZsRE from English to Chinese. Then, we conduct English editing on various knowledge editing methods covering different paradigms, and evaluate their performance in Chinese, and vice versa. To give deeper analyses of the crosslingual effect, the evaluation includes four aspects, i.e., reliability, generality, locality and portability. Furthermore, we analyze the inconsistent behaviors of the edited models and discuss their specific challenges.
引用
收藏
页码:11676 / 11686
页数:11
相关论文
共 12 条
  • [1] Dai DM, 2022, PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), P8493
  • [2] De Cao N, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P6491
  • [3] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [4] RETRACTED: Quality Evaluation Model of Education in Remote Learning and IoT Based on College Students' Quality Cultivation (Retracted Article)
    Huang, Yawen
    Wang, Jing
    [J]. WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [5] Natural Questions: A Benchmark for Question Answering Research
    Kwiatkowski, Tom
    Palomaki, Jennimaria
    Redfield, Olivia
    Collins, Michael
    Parikh, Ankur
    Alberti, Chris
    Epstein, Danielle
    Polosukhin, Illia
    Devlin, Jacob
    Lee, Kenton
    Toutanova, Kristina
    Jones, Llion
    Kelcey, Matthew
    Chang, Ming-Wei
    Dai, Andrew M.
    Uszkoreit, Jakob
    Quoc Le
    Petrov, Slav
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2019, 7 : 453 - 466
  • [6] Liang YL, 2023, PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, P2934
  • [7] Meng Kevin, 2022, Advances in Neural Information Processing Systems, V35, P17359
  • [8] Rajpurkar P., 2016, P 2016 C EMP METH NA, P2383, DOI [DOI 10.18653/V1/D16-1264, 10.18653/v1/D16- 1264]
  • [9] Wang JA, 2023, PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, P15127
  • [10] A Survey on Cross-Lingual Summarization
    Wang, Jiaan
    Meng, Fandong
    Zheng, Duo
    Liang, Yunlong
    Li, Zhixu
    Qu, Jianfeng
    Zhou, Jie
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1304 - 1323