Large language models in medicine
Large language models (LLMs) can respond to free-text queries without being specifically
trained in the task in question, causing excitement and concern about their use in healthcare …
trained in the task in question, causing excitement and concern about their use in healthcare …
Challenges and applications of large language models
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …
A survey of large language models
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
Scaling data-constrained language models
The current trend of scaling language models involves increasing both parameter count and
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …
The RefinedWeb dataset for Falcon LLM: Outperforming curated corpora with web data only
Large language models are commonly trained on a mixture of filtered web data and
curated``high-quality''corpora, such as social media conversations, books, or technical …
curated``high-quality''corpora, such as social media conversations, books, or technical …
A pretrainer's guide to training data: Measuring the effects of data age, domain coverage, quality, & toxicity
Pretraining is the preliminary and fundamental step in developing capable language models
(LM). Despite this, pretraining data design is critically under-documented and often guided …
(LM). Despite this, pretraining data design is critically under-documented and often guided …
Red teaming chatgpt via jailbreaking: Bias, robustness, reliability and toxicity
Recent breakthroughs in natural language processing (NLP) have permitted the synthesis
and comprehension of coherent text in an open-ended way, therefore translating the …
and comprehension of coherent text in an open-ended way, therefore translating the …
Self-consuming generative models go mad
Seismic advances in generative AI algorithms for imagery, text, and other data types has led
to the temptation to use synthetic data to train next-generation models. Repeating this …
to the temptation to use synthetic data to train next-generation models. Repeating this …
To repeat or not to repeat: Insights from scaling llm under token-crisis
Recent research has highlighted the importance of dataset size in scaling language models.
However, large language models (LLMs) are notoriously token-hungry during pre-training …
However, large language models (LLMs) are notoriously token-hungry during pre-training …
When foundation model meets federated learning: Motivations, challenges, and future directions
The intersection of the Foundation Model (FM) and Federated Learning (FL) provides mutual
benefits, presents a unique opportunity to unlock new possibilities in AI research, and …
benefits, presents a unique opportunity to unlock new possibilities in AI research, and …