Parameter-efficient fine-tuning for large models: A comprehensive survey
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …
enabling remarkable achievements across various tasks. However, their unprecedented …
Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …
language models (PLMs), particularly the emergence of large language models (LLMs) with …
MELoRA: mini-ensemble low-rank adapters for parameter-efficient fine-tuning
Parameter-efficient fine-tuning (PEFT) is a popular method for tailoring pre-trained large
language models (LLMs), especially as the models' scale and the diversity of tasks increase …
language models (LLMs), especially as the models' scale and the diversity of tasks increase …
Increlora: Incremental parameter allocation method for parameter-efficient fine-tuning
With the increasing size of pre-trained language models (PLMs), fine-tuning all the
parameters in the model is not efficient, especially when there are a large number of …
parameters in the model is not efficient, especially when there are a large number of …
Survey of different large language model architectures: Trends, benchmarks, and challenges
Large Language Models (LLMs) represent a class of deep learning models adept at
understanding natural language and generating coherent text in response to prompts or …
understanding natural language and generating coherent text in response to prompts or …
Low-rank adaptation of large language model rescoring for parameter-efficient speech recognition
We propose a neural language modeling system based on low-rank adaptation (LoRA) for
speech recognition output rescoring. Although pretrained language models (LMs) like BERT …
speech recognition output rescoring. Although pretrained language models (LMs) like BERT …
Pissa: Principal singular values and singular vectors adaptation of large language models
As the parameters of LLMs expand, the computational cost of fine-tuning the entire model
becomes prohibitive. To address this challenge, we introduce a PEFT method, Principal …
becomes prohibitive. To address this challenge, we introduce a PEFT method, Principal …
RST-LoRA: A Discourse-Aware Low-Rank Adaptation for Long Document Abstractive Summarization
D Pu, V Demberg - arXiv preprint arXiv:2405.00657, 2024 - arxiv.org
For long document summarization, discourse structure is important to discern the key
content of the text and the differences in importance level between sentences. Unfortunately …
content of the text and the differences in importance level between sentences. Unfortunately …
Personalized Federated Instruction Tuning via Neural Architecture Search
P Zhang, Y Zhou, M Hu, J Feng, J Weng… - arXiv preprint arXiv …, 2024 - arxiv.org
Federated Instruction Tuning (FIT) has shown the ability to achieve collaborative model
instruction tuning among massive data owners without sharing private data. However, it still …
instruction tuning among massive data owners without sharing private data. However, it still …
Memory-Efficient Fine-Tuning of Transformers via Token Selection
Fine-tuning provides an effective means to specialize pre-trained models for various
downstream tasks. However, fine-tuning often incurs high memory overhead, especially for …
downstream tasks. However, fine-tuning often incurs high memory overhead, especially for …