Qlora: Efficient finetuning of quantized llms
We present QLoRA, an efficient finetuning approach that reduces memory usage enough to
finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit …
finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit …
Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …
previously-unseen task without any gradient-based training by feeding a small number of …
Rlprompt: Optimizing discrete text prompts with reinforcement learning
Prompting has shown impressive success in enabling large pretrained language models
(LMs) to perform diverse NLP tasks, especially when only few downstream data are …
(LMs) to perform diverse NLP tasks, especially when only few downstream data are …
“What it wants me to say”: Bridging the abstraction gap between end-user programmers and code-generating large language models
Code-generating large language models map natural language to code. However, only a
small portion of the infinite space of naturalistic utterances is effective at guiding code …
small portion of the infinite space of naturalistic utterances is effective at guiding code …
Guiding large language models via directional stimulus prompting
Abstract We introduce Directional Stimulus Prompting, a novel framework for guiding black-
box large language models (LLMs) towards specific desired outputs. Instead of directly …
box large language models (LLMs) towards specific desired outputs. Instead of directly …
Lorahub: Efficient cross-task generalization via dynamic lora composition
Low-rank adaptations (LoRA) are often employed to fine-tune large language models
(LLMs) for new tasks. This paper investigates LoRA composability for cross-task …
(LLMs) for new tasks. This paper investigates LoRA composability for cross-task …
Overcoming catastrophic forgetting in zero-shot cross-lingual generation
In this paper, we explore the challenging problem of performing a generative task in a target
language when labeled data is only available in English, using summarization as a case …
language when labeled data is only available in English, using summarization as a case …
Aprompt: Attention prompt tuning for efficient adaptation of pre-trained language models
With the continuous growth of large language models, the process of fine-tuning these
models for new tasks has become increasingly parameter-intensive. Prompt tuning, a …
models for new tasks has become increasingly parameter-intensive. Prompt tuning, a …
Preference-grounded token-level guidance for language model fine-tuning
Aligning language models (LMs) with preferences is an important problem in natural
language generation. A key challenge is that preferences are typically provided at the …
language generation. A key challenge is that preferences are typically provided at the …
How does in-context learning help prompt tuning?
Fine-tuning large language models is becoming ever more impractical due to their rapidly-
growing scale. This motivates the use of parameter-efficient adaptation methods such as …
growing scale. This motivates the use of parameter-efficient adaptation methods such as …