Augmented language models: a survey
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …
skills and the ability to use tools. The former is defined as decomposing a potentially …
One embedder, any task: Instruction-finetuned text embeddings
We introduce INSTRUCTOR, a new method for computing text embeddings given task
instructions: every text input is embedded together with instructions explaining the use case …
instructions: every text input is embedded together with instructions explaining the use case …
Recommendation as instruction following: A large language model empowered recommendation approach
In the past decades, recommender systems have attracted much attention in both research
and industry communities, and a large number of studies have been devoted to developing …
and industry communities, and a large number of studies have been devoted to developing …
Large language models for information retrieval: A survey
As a primary means of information acquisition, information retrieval (IR) systems, such as
search engines, have integrated themselves into our daily lives. These systems also serve …
search engines, have integrated themselves into our daily lives. These systems also serve …
Large language models are effective text rankers with pairwise ranking prompting
Ranking documents using Large Language Models (LLMs) by directly feeding the query and
candidate documents into the prompt is an interesting and practical problem. However …
candidate documents into the prompt is an interesting and practical problem. However …
C-pack: Packaged resources to advance general chinese embedding
We introduce C-Pack, a package of resources that significantly advance the field of general
Chinese embeddings. C-Pack includes three critical resources. 1) C-MTEB is a …
Chinese embeddings. C-Pack includes three critical resources. 1) C-MTEB is a …
Enhancing retrieval-augmented large language models with iterative retrieval-generation synergy
Large language models are powerful text processors and reasoners, but are still subject to
limitations including outdated knowledge and hallucinations, which necessitates connecting …
limitations including outdated knowledge and hallucinations, which necessitates connecting …
Exploring the benefits of training expert language models over instruction tuning
Abstract Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …
as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen …
Precise zero-shot dense retrieval without relevance labels
While dense retrieval has been shown effective and efficient across tasks and languages, it
remains difficult to create effective fully zero-shot dense retrieval systems when no relevance …
remains difficult to create effective fully zero-shot dense retrieval systems when no relevance …
How to train your dragon: Diverse augmentation towards generalizable dense retrieval
Various techniques have been developed in recent years to improve dense retrieval (DR),
such as unsupervised contrastive learning and pseudo-query generation. Existing DRs …
such as unsupervised contrastive learning and pseudo-query generation. Existing DRs …