arXiv Analytics

Sign in

arXiv:2311.09632 [cs.CL]AbstractReferencesReviewsResources

Online Continual Knowledge Learning for Language Models

Yuhao Wu, Tongjun Shi, Karthick Sharma, Chun Wei Seah, Shuhao Zhang

Published 2023-11-16Version 1

Large Language Models (LLMs) serve as repositories of extensive world knowledge, enabling them to perform tasks such as question-answering and fact-checking. However, this knowledge can become obsolete as global contexts change. In this paper, we introduce a novel problem in the realm of continual learning: Online Continual Knowledge Learning (OCKL). This problem formulation aims to manage the dynamic nature of world knowledge in LMs under real-time constraints. We propose a new benchmark and evaluation metric designed to measure both the rate of new knowledge acquisition and the retention of previously learned knowledge. Our empirical evaluation, conducted using a variety of state-of-the-art methods, establishes robust base-lines for OCKL. Our results reveal that existing continual learning approaches are unfortunately insufficient for tackling the unique challenges posed by OCKL. We identify key factors that influence the trade-off between knowledge acquisition and retention, thereby advancing our understanding of how to train LMs in a continually evolving environment.

Related articles: Most relevant | Search more
arXiv:2202.00828 [cs.CL] (Published 2022-02-02)
Co-training Improves Prompt-based Learning for Large Language Models
arXiv:2205.11916 [cs.CL] (Published 2022-05-24)
Large Language Models are Zero-Shot Reasoners
arXiv:2102.07350 [cs.CL] (Published 2021-02-15)
Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm