Qlora: Efficient finetuning of quantized llms

T Dettmers, A Pagnoni, A Holtzman… - Advances in Neural …, 2024 - proceedings.neurips.cc
We present QLoRA, an efficient finetuning approach that reduces memory usage enough to
finetune a 65B parameter model on a single 48GB GPU while preserving full 16-bit …

Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning

H Liu, D Tam, M Muqeeth, J Mohta… - Advances in …, 2022 - proceedings.neurips.cc
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …

Rlprompt: Optimizing discrete text prompts with reinforcement learning

M Deng, J Wang, CP Hsieh, Y Wang, H Guo… - arXiv preprint arXiv …, 2022 - arxiv.org
Prompting has shown impressive success in enabling large pretrained language models
(LMs) to perform diverse NLP tasks, especially when only few downstream data are …

“What it wants me to say”: Bridging the abstraction gap between end-user programmers and code-generating large language models

MX Liu, A Sarkar, C Negreanu, B Zorn… - Proceedings of the …, 2023 - dl.acm.org
Code-generating large language models map natural language to code. However, only a
small portion of the infinite space of naturalistic utterances is effective at guiding code …

Guiding large language models via directional stimulus prompting

Z Li, B Peng, P He, M Galley… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract We introduce Directional Stimulus Prompting, a novel framework for guiding black-
box large language models (LLMs) towards specific desired outputs. Instead of directly …

Lorahub: Efficient cross-task generalization via dynamic lora composition

C Huang, Q Liu, BY Lin, T Pang, C Du, M Lin - arXiv preprint arXiv …, 2023 - arxiv.org
Low-rank adaptations (LoRA) are often employed to fine-tune large language models
(LLMs) for new tasks. This paper investigates LoRA composability for cross-task …

Overcoming catastrophic forgetting in zero-shot cross-lingual generation

T Vu, A Barua, B Lester, D Cer, M Iyyer… - arXiv preprint arXiv …, 2022 - arxiv.org
In this paper, we explore the challenging problem of performing a generative task in a target
language when labeled data is only available in English, using summarization as a case …

Aprompt: Attention prompt tuning for efficient adaptation of pre-trained language models

Q Wang, Y Mao, J Wang, H Yu, S Nie… - Proceedings of the …, 2023 - aclanthology.org
With the continuous growth of large language models, the process of fine-tuning these
models for new tasks has become increasingly parameter-intensive. Prompt tuning, a …

Preference-grounded token-level guidance for language model fine-tuning

S Yang, S Zhang, C Xia, Y Feng… - Advances in Neural …, 2024 - proceedings.neurips.cc
Aligning language models (LMs) with preferences is an important problem in natural
language generation. A key challenge is that preferences are typically provided at the …

How does in-context learning help prompt tuning?

S Sun, Y Liu, D Iter, C Zhu, M Iyyer - arXiv preprint arXiv:2302.11521, 2023 - arxiv.org
Fine-tuning large language models is becoming ever more impractical due to their rapidly-
growing scale. This motivates the use of parameter-efficient adaptation methods such as …