Understanding llms: A comprehensive overview from training to inference

Y Liu, H He, T Han, X Zhang, M Liu, J Tian… - arXiv preprint arXiv …, 2024 - arxiv.org
The introduction of ChatGPT has led to a significant increase in the utilization of Large
Language Models (LLMs) for addressing downstream tasks. There's an increasing focus on …

Datasets for large language models: A comprehensive survey

Y Liu, J Cao, C Liu, K Ding, L Jin - arXiv preprint arXiv:2402.18041, 2024 - arxiv.org
This paper embarks on an exploration into the Large Language Model (LLM) datasets,
which play a crucial role in the remarkable advancements of LLMs. The datasets serve as …

Starcoder 2 and the stack v2: The next generation

A Lozhkov, R Li, LB Allal, F Cassano… - arXiv preprint arXiv …, 2024 - arxiv.org
The BigCode project, an open-scientific collaboration focused on the responsible
development of Large Language Models for Code (Code LLMs), introduces StarCoder2. In …

Show-o: One single transformer to unify multimodal understanding and generation

J Xie, W Mao, Z Bai, DJ Zhang, W Wang, KQ Lin… - arXiv preprint arXiv …, 2024 - arxiv.org
We present a unified transformer, ie, Show-o, that unifies multimodal understanding and
generation. Unlike fully autoregressive models, Show-o unifies autoregressive and …

Building and better understanding vision-language models: insights and future directions

H Laurençon, A Marafioti, V Sanh… - arXiv preprint arXiv …, 2024 - arxiv.org
The field of vision-language models (VLMs), which take images and texts as inputs and
output texts, is rapidly evolving and has yet to reach consensus on several key aspects of …

Data management for large language models: A survey

Z Wang, W Zhong, Y Wang, Q Zhu, F Mi… - arXiv preprint arXiv …, 2023 - arxiv.org
Data plays a fundamental role in the training of Large Language Models (LLMs). Effective
data management, particularly in the formulation of a well-suited training dataset, holds …

How to train long-context language models (effectively)

T Gao, A Wettig, H Yen, D Chen - arXiv preprint arXiv:2410.02660, 2024 - arxiv.org
We study continued training and supervised fine-tuning (SFT) of a language model (LM) to
make effective use of long-context information. We first establish a reliable evaluation …

Qurating: Selecting high-quality data for training language models

A Wettig, A Gupta, S Malik, D Chen - arXiv preprint arXiv:2402.09739, 2024 - arxiv.org
Selecting high-quality pre-training data is important for creating capable language models,
but existing methods rely on simple heuristics. We introduce QuRating, a method for …

Resolving discrepancies in compute-optimal scaling of language models

T Porian, M Wortsman, J Jitsev, L Schmidt… - arXiv preprint arXiv …, 2024 - arxiv.org
Kaplan et al. and Hoffmann et al. developed influential scaling laws for the optimal model
size as a function of the compute budget, but these laws yield substantially different …

From matching to generation: A survey on generative information retrieval

X Li, J Jin, Y Zhou, Y Zhang, P Zhang, Y Zhu… - arXiv preprint arXiv …, 2024 - arxiv.org
Information Retrieval (IR) systems are crucial tools for users to access information, widely
applied in scenarios like search engines, question answering, and recommendation …