Give me the facts! a survey on factual knowledge probing in pre-trained language models

P Youssef, OA Koraş, M Li, J Schlötterer… - arXiv preprint arXiv …, 2023 - arxiv.org
Pre-trained Language Models (PLMs) are trained on vast unlabeled data, rich in world
knowledge. This fact has sparked the interest of the community in quantifying the amount of …

Carpe diem: on the evaluation of world knowledge in lifelong language models

Y Kim, J Yoon, S Ye, SJ Hwang, S Yun - arXiv preprint arXiv:2311.08106, 2023 - arxiv.org
In an ever-evolving world, the dynamic nature of knowledge presents challenges for
language models that are trained on static data, leading to outdated encoded information …

Bitimebert: Extending pre-trained language representations with bi-temporal information

J Wang, A Jatowt, M Yoshikawa, Y Cai - Proceedings of the 46th …, 2023 - dl.acm.org
Time is an important aspect of documents and is used in a range of NLP and IR tasks. In this
work, we investigate methods for incorporating temporal information during pre-training to …

Towards Effective Time-Aware Language Representation: Exploring Enhanced Temporal Understanding in Language Models

J Wang, A Jatowt, Y Cai - arXiv preprint arXiv:2406.01863, 2024 - arxiv.org
In the evolving field of Natural Language Processing, understanding the temporal context of
text is increasingly crucial. This study investigates methods to incorporate temporal …

Difference-masking: Choosing what to mask in continued pretraining

A Wilf, SN Akter, L Mathur, PP Liang, S Mathew… - arXiv preprint arXiv …, 2023 - arxiv.org
The self-supervised objective of masking-and-predicting has led to promising performance
gains on a variety of downstream tasks. However, while most approaches randomly mask …

HistoChatbot: Educating History by Generating Quizzes in Social Network Services

Y Sumikawa, A Jatowt - International Conference on Asian Digital Libraries, 2023 - Springer
Microblogging platforms can provide novel, attractive opportunities for communicating and
disseminating content about important events from the past. We propose a novel framework …