A comprehensive survey of continual learning: theory, method and application
To cope with real-world dynamics, an intelligent system needs to incrementally acquire,
update, accumulate, and exploit knowledge throughout its lifetime. This ability, known as …
update, accumulate, and exploit knowledge throughout its lifetime. This ability, known as …
Meditron-70b: Scaling medical pretraining for large language models
Large language models (LLMs) can potentially democratize access to medical knowledge.
While many efforts have been made to harness and improve LLMs' medical knowledge and …
While many efforts have been made to harness and improve LLMs' medical knowledge and …
Lfpt5: A unified framework for lifelong few-shot language learning based on prompt tuning of t5
Existing approaches to lifelong language learning rely on plenty of labeled data for learning
a new task, which is hard to obtain in most real scenarios. Considering that humans can …
a new task, which is hard to obtain in most real scenarios. Considering that humans can …
Computational models to study language processing in the human brain: A survey
Despite differing from the human language processing mechanism in implementation and
algorithms, current language models demonstrate remarkable human-like or surpassing …
algorithms, current language models demonstrate remarkable human-like or surpassing …
Continual sequence generation with adaptive compositional modules
Continual learning is essential for real-world deployment when there is a need to quickly
adapt the model to new tasks without forgetting knowledge of old tasks. Existing work on …
adapt the model to new tasks without forgetting knowledge of old tasks. Existing work on …
Contintin: Continual learning from task instructions
The mainstream machine learning paradigms for NLP often work with two underlying
presumptions. First, the target task is predefined and static; a system merely needs to learn …
presumptions. First, the target task is predefined and static; a system merely needs to learn …
Incremental prompting: Episodic memory prompt for lifelong event detection
Lifelong event detection aims to incrementally update a model with new event types and
data while retaining the capability on previously learned old types. One critical challenge is …
data while retaining the capability on previously learned old types. One critical challenge is …
Fine-tuned vs. prompt-tuned supervised representations: Which better account for brain language representations?
To decipher the algorithm underlying the human brain's language representation, previous
work probed brain responses to language input with pre-trained artificial neural network …
work probed brain responses to language input with pre-trained artificial neural network …
Towards quantifiable dialogue coherence evaluation
Automatic dialogue coherence evaluation has attracted increasing attention and is crucial
for developing promising dialogue systems. However, existing metrics have two major …
for developing promising dialogue systems. However, existing metrics have two major …
Saullm-54b & saullm-141b: Scaling up domain adaptation for the legal domain
In this paper, we introduce SaulLM-medium and SaulLM-large, two large language models
(LLMs) families tailored for the legal sector. These models, which feature architectures of 54 …
(LLMs) families tailored for the legal sector. These models, which feature architectures of 54 …