Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arXiv preprint arXiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment

L Xu, H Xie, SZJ Qin, X Tao, FL Wang - arXiv preprint arXiv:2312.12148, 2023 - arxiv.org
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …

MELoRA: mini-ensemble low-rank adapters for parameter-efficient fine-tuning

P Ren, C Shi, S Wu, M Zhang, Z Ren… - Proceedings of the …, 2024 - aclanthology.org
Parameter-efficient fine-tuning (PEFT) is a popular method for tailoring pre-trained large
language models (LLMs), especially as the models' scale and the diversity of tasks increase …

Increlora: Incremental parameter allocation method for parameter-efficient fine-tuning

F Zhang, L Li, J Chen, Z Jiang, B Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
With the increasing size of pre-trained language models (PLMs), fine-tuning all the
parameters in the model is not efficient, especially when there are a large number of …

Survey of different large language model architectures: Trends, benchmarks, and challenges

M Shao, A Basit, R Karri, M Shafique - IEEE Access, 2024 - ieeexplore.ieee.org
Large Language Models (LLMs) represent a class of deep learning models adept at
understanding natural language and generating coherent text in response to prompts or …

Low-rank adaptation of large language model rescoring for parameter-efficient speech recognition

Y Yu, CHH Yang, J Kolehmainen… - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
We propose a neural language modeling system based on low-rank adaptation (LoRA) for
speech recognition output rescoring. Although pretrained language models (LMs) like BERT …

Pissa: Principal singular values and singular vectors adaptation of large language models

F Meng, Z Wang, M Zhang - arXiv preprint arXiv:2404.02948, 2024 - arxiv.org
As the parameters of LLMs expand, the computational cost of fine-tuning the entire model
becomes prohibitive. To address this challenge, we introduce a PEFT method, Principal …

RST-LoRA: A Discourse-Aware Low-Rank Adaptation for Long Document Abstractive Summarization

D Pu, V Demberg - arXiv preprint arXiv:2405.00657, 2024 - arxiv.org
For long document summarization, discourse structure is important to discern the key
content of the text and the differences in importance level between sentences. Unfortunately …

Personalized Federated Instruction Tuning via Neural Architecture Search

P Zhang, Y Zhou, M Hu, J Feng, J Weng… - arXiv preprint arXiv …, 2024 - arxiv.org
Federated Instruction Tuning (FIT) has shown the ability to achieve collaborative model
instruction tuning among massive data owners without sharing private data. However, it still …

Memory-Efficient Fine-Tuning of Transformers via Token Selection

A Simoulin, N Park, X Liu, G Yang - Proceedings of the 2024 …, 2024 - aclanthology.org
Fine-tuning provides an effective means to specialize pre-trained models for various
downstream tasks. However, fine-tuning often incurs high memory overhead, especially for …