Large language models suffer from their own output: An analysis of the self-consuming training loop

M Briesch, D Sobania, F Rothlauf - arXiv preprint arXiv:2311.16822, 2023 - arxiv.org
Large language models (LLM) have become state of the art in many benchmarks and
conversational LLM applications like ChatGPT are now widely used by the public. Those …

A tale of tails: Model collapse as a change of scaling laws

E Dohmatob, Y Feng, P Yang, F Charton… - arXiv preprint arXiv …, 2024 - arxiv.org
As AI model size grows, neural scaling laws have become a crucial tool to predict the
improvements of large models when increasing capacity and the size of original (human or …

Beware of words: Evaluating the lexical diversity of conversational llms using chatgpt as case study

G Martínez, JA Hernández, J Conde… - ACM Transactions on …, 2024 - dl.acm.org
The performance of conversational Large Language Models (LLMs) in general, and of
ChatGPT in particular, is currently being evaluated on many different tasks, from logical …

Beyond model collapse: Scaling up with synthesized data requires reinforcement

Y Feng, E Dohmatob, P Yang, F Charton… - ICML 2024 Workshop …, 2024 - openreview.net
Synthesized data from generative models is increasingly considered as an alternative to
human-annotated data for fine-tuning Large Language Models. This raises concerns about …

A survey on the impact of AI-based recommenders on human behaviours: methodologies, outcomes and future directions

L Pappalardo, E Ferragina, S Citraro… - arXiv preprint arXiv …, 2024 - arxiv.org
Recommendation systems and assistants (in short, recommenders) are ubiquitous in online
platforms and influence most actions of our day-to-day lives, suggesting items or providing …

Self-consuming generative models with curated data provably optimize human preferences

D Ferbach, Q Bertrand, AJ Bose, G Gidel - arXiv preprint arXiv:2407.09499, 2024 - arxiv.org
The rapid progress in generative models has resulted in impressive leaps in generation
quality, blurring the lines between synthetic and real data. Web-scale datasets are now …

Strong model collapse

E Dohmatob, Y Feng, A Subramonian… - arXiv preprint arXiv …, 2024 - arxiv.org
Within the scaling laws paradigm, which underpins the training of large neural networks like
ChatGPT and Llama, we consider a supervised regression setting and establish the …

Human vs. Generative AI in Content Creation Competition: Symbiosis or Conflict?

F Yao, C Li, D Nekipelov, H Wang, H Xu - arXiv preprint arXiv:2402.15467, 2024 - arxiv.org
The advent of generative AI (GenAI) technology produces transformative impact on the
content creation landscape, offering alternative approaches to produce diverse, high-quality …

Model collapse demystified: The case of regression

E Dohmatob, Y Feng, J Kempe - arXiv preprint arXiv:2402.07712, 2024 - arxiv.org
In the era of large language models like ChatGPT, the phenomenon of" model collapse"
refers to the situation whereby as a model is trained recursively on data generated from …

Regurgitative training: The value of real data in training large language models

J Zhang, D Qiao, M Yang, Q Wei - arXiv preprint arXiv:2407.12835, 2024 - arxiv.org
What happens if we train a new Large Language Model (LLM) using data that are at least
partially generated by other LLMs? The explosive success of LLMs means that a substantial …