Augmented language models: a survey

G Mialon, R Dessì, M Lomeli, C Nalmpantis… - arXiv preprint arXiv …, 2023 - arxiv.org
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …

One embedder, any task: Instruction-finetuned text embeddings

H Su, W Shi, J Kasai, Y Wang, Y Hu… - arXiv preprint arXiv …, 2022 - arxiv.org
We introduce INSTRUCTOR, a new method for computing text embeddings given task
instructions: every text input is embedded together with instructions explaining the use case …

Recommendation as instruction following: A large language model empowered recommendation approach

J Zhang, R Xie, Y Hou, WX Zhao, L Lin… - arXiv preprint arXiv …, 2023 - arxiv.org
In the past decades, recommender systems have attracted much attention in both research
and industry communities, and a large number of studies have been devoted to developing …

Large language models for information retrieval: A survey

Y Zhu, H Yuan, S Wang, J Liu, W Liu, C Deng… - arXiv preprint arXiv …, 2023 - arxiv.org
As a primary means of information acquisition, information retrieval (IR) systems, such as
search engines, have integrated themselves into our daily lives. These systems also serve …

Large language models are effective text rankers with pairwise ranking prompting

Z Qin, R Jagerman, K Hui, H Zhuang, J Wu… - arXiv preprint arXiv …, 2023 - arxiv.org
Ranking documents using Large Language Models (LLMs) by directly feeding the query and
candidate documents into the prompt is an interesting and practical problem. However …

C-pack: Packaged resources to advance general chinese embedding

S Xiao, Z Liu, P Zhang, N Muennighof - arXiv preprint arXiv:2309.07597, 2023 - arxiv.org
We introduce C-Pack, a package of resources that significantly advance the field of general
Chinese embeddings. C-Pack includes three critical resources. 1) C-MTEB is a …

Enhancing retrieval-augmented large language models with iterative retrieval-generation synergy

Z Shao, Y Gong, Y Shen, M Huang, N Duan… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models are powerful text processors and reasoners, but are still subject to
limitations including outdated knowledge and hallucinations, which necessitates connecting …

Exploring the benefits of training expert language models over instruction tuning

J Jang, S Kim, S Ye, D Kim… - International …, 2023 - proceedings.mlr.press
Abstract Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …

Precise zero-shot dense retrieval without relevance labels

L Gao, X Ma, J Lin, J Callan - arXiv preprint arXiv:2212.10496, 2022 - arxiv.org
While dense retrieval has been shown effective and efficient across tasks and languages, it
remains difficult to create effective fully zero-shot dense retrieval systems when no relevance …

How to train your dragon: Diverse augmentation towards generalizable dense retrieval

SC Lin, A Asai, M Li, B Oguz, J Lin, Y Mehdad… - arXiv preprint arXiv …, 2023 - arxiv.org
Various techniques have been developed in recent years to improve dense retrieval (DR),
such as unsupervised contrastive learning and pseudo-query generation. Existing DRs …