Pre-trained language models for text generation: A survey
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …
Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …
language models (PLMs), particularly the emergence of large language models (LLMs) with …
A survey of large language models
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
Llama-adapter: Efficient fine-tuning of language models with zero-init attention
We present LLaMA-Adapter, a lightweight adaption method to efficiently fine-tune LLaMA
into an instruction-following model. Using 52K self-instruct demonstrations, LLaMA-Adapter …
into an instruction-following model. Using 52K self-instruct demonstrations, LLaMA-Adapter …
Aligning large language models with human: A survey
Large Language Models (LLMs) trained on extensive textual corpora have emerged as
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …
A survey on model compression for large language models
Large Language Models (LLMs) have revolutionized natural language processing tasks with
remarkable success. However, their formidable size and computational demands present …
remarkable success. However, their formidable size and computational demands present …
Loftq: Lora-fine-tuning-aware quantization for large language models
Quantization is an indispensable technique for serving Large Language Models (LLMs) and
has recently found its way into LoRA fine-tuning. In this work we focus on the scenario where …
has recently found its way into LoRA fine-tuning. In this work we focus on the scenario where …
[HTML][HTML] Information retrieval meets large language models: a strategic report from chinese ir community
The research field of Information Retrieval (IR) has evolved significantly, expanding beyond
traditional search to meet diverse user information needs. Recently, Large Language …
traditional search to meet diverse user information needs. Recently, Large Language …
A simple recipe for contrastively pre-training video-first encoders beyond 16 frames
Understanding long real-world videos requires modeling of long-range visual
dependencies. To this end we explore video-first architectures building on the common …
dependencies. To this end we explore video-first architectures building on the common …
One-for-all: Generalized lora for parameter-efficient fine-tuning
We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-
efficient fine-tuning tasks. Enhancing Low-Rank Adaptation (LoRA), GLoRA employs a …
efficient fine-tuning tasks. Enhancing Low-Rank Adaptation (LoRA), GLoRA employs a …