Time-Aware Language Models as Temporal Knowledge Bases

被引:58
|
作者
Dhingra, Bhuwan [1 ,2 ]
Cole, Jeremy R. R. [1 ]
Eisenschlos, Julian Martin [1 ]
Gillick, Daniel [1 ]
Eisenstein, Jacob [1 ]
Cohen, William W. W. [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
[2] Duke Univ, Durham, NC 27706 USA
关键词
Compendex;
D O I
10.1162/tacl_a_00459
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many facts come with an expiration date, from the name of the President to the basketball team Lebron James plays for. However, most language models (LMs) are trained on snapshots of data collected at a specific moment in time. This can limit their utility, especially in the closed-book setting where the pretraining corpus must contain the facts the model should memorize. We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum-those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data. To mitigate these problems, we propose a simple technique for jointly modeling text with its timestamp. This improves memorization of seen facts from the training time period, as well as calibration on predictions about unseen facts from future time periods. We also show that models trained with temporal context can be efficiently "refreshed" as new data arrives, without the need for retraining from scratch.
引用
收藏
页码:257 / 273
页数:17
相关论文
共 5 条
  • [1] Time-aware Random Walk Diffusion to Improve Dynamic Graph Learning
    Lee, Jong-Whi
    Jung, Jinhong
    arXiv, 2022,
  • [2] TrojBits: A Hardware Aware Inference-Time Attack on Transformer-Based Language Models
    Al Ghanim, Mansour
    Santriaji, Muhammad
    Lou, Qian
    Solihin, Yan
    Frontiers in Artificial Intelligence and Applications, 2023, 372 : 60 - 68
  • [3] Camg: Context-Aware Moment Graph Network for Multimodal Temporal Activity  Localization Via Language
    Hu, Yuelin
    Xu, Yuanwu
    Zhang, Yuejie
    Feng, Rui
    Zhang, Tao
    Lu, Xuequan
    Gao, Shang
    SSRN, 2022,
  • [4] CommonIT: Commonality-Aware Instruction Tuning for Large Language Models via Data Partitions
    Rao, Jun
    Liu, Xuebo
    Lian, Lian
    Cheng, Shengjun
    Liao, Yunjie
    Zhang, Min
    EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference, 2024, : 10064 - 10083
  • [5] Multi-modal temporal attention models for crop mapping from satellite time series
    Sainte Fare Garnot, Vivien
    Landrieu, Loic
    Chehata, Nesrine
    ISPRS Journal of Photogrammetry and Remote Sensing, 2022, 187 : 294 - 305