Domain specialization as the key to make large language models disruptive: A comprehensive survey
Large language models (LLMs) have significantly advanced the field of natural language
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …
Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …
language models (PLMs), particularly the emergence of large language models (LLMs) with …
Vl-adapter: Parameter-efficient transfer learning for vision-and-language tasks
Recently, fine-tuning language models pre-trained on large text corpora have provided huge
improvements on vision-and-language (V&L) tasks as well as on pure language tasks …
improvements on vision-and-language (V&L) tasks as well as on pure language tasks …
Compacter: Efficient low-rank hypercomplex adapter layers
R Karimi Mahabadi, J Henderson… - Advances in Neural …, 2021 - proceedings.neurips.cc
Adapting large-scale pretrained language models to downstream tasks via fine-tuning is the
standard method for achieving state-of-the-art performance on NLP benchmarks. However …
standard method for achieving state-of-the-art performance on NLP benchmarks. However …
Dive into deep learning
This open-source book represents our attempt to make deep learning approachable,
teaching readers the concepts, the context, and the code. The entire book is drafted in …
teaching readers the concepts, the context, and the code. The entire book is drafted in …
Scale efficiently: Insights from pre-training and fine-tuning transformers
There remain many open questions pertaining to the scaling behaviour of Transformer
architectures. These scaling decisions and findings can be critical, as training runs often …
architectures. These scaling decisions and findings can be critical, as training runs often …
Interactive natural language processing
Interactive Natural Language Processing (iNLP) has emerged as a novel paradigm within
the field of NLP, aimed at addressing limitations in existing frameworks while aligning with …
the field of NLP, aimed at addressing limitations in existing frameworks while aligning with …
Polyhistor: Parameter-efficient multi-task adaptation for dense vision tasks
Adapting large-scale pretrained models to various downstream tasks via fine-tuning is a
standard method in machine learning. Recently, parameter-efficient fine-tuning methods …
standard method in machine learning. Recently, parameter-efficient fine-tuning methods …
Perfect: Prompt-free and efficient few-shot learning with language models
RK Mahabadi, L Zettlemoyer, J Henderson… - arXiv preprint arXiv …, 2022 - arxiv.org
Current methods for few-shot fine-tuning of pretrained masked language models (PLMs)
require carefully engineered prompts and verbalizers for each new task to convert examples …
require carefully engineered prompts and verbalizers for each new task to convert examples …