Instruction tuning for large language models: A survey
This paper surveys research works in the quickly advancing field of instruction tuning (IT), a
crucial technique to enhance the capabilities and controllability of large language models …
crucial technique to enhance the capabilities and controllability of large language models …
Large ai models in health informatics: Applications, challenges, and the future
Large AI models, or foundation models, are models recently emerging with massive scales
both parameter-wise and data-wise, the magnitudes of which can reach beyond billions …
both parameter-wise and data-wise, the magnitudes of which can reach beyond billions …
A survey of large language models
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …
Enhancing chat language models by scaling high-quality instructional conversations
Fine-tuning on instruction data has been widely validated as an effective practice for
implementing chat language models like ChatGPT. Scaling the diversity and quality of such …
implementing chat language models like ChatGPT. Scaling the diversity and quality of such …
Language models are super mario: Absorbing abilities from homologous models as a free lunch
In this paper, we unveil that Language Models (LMs) can acquire new capabilities by
assimilating parameters from homologous models without retraining or GPUs. We first …
assimilating parameters from homologous models without retraining or GPUs. We first …
A survey on model compression for large language models
Large Language Models (LLMs) have revolutionized natural language processing tasks with
remarkable success. However, their formidable size and computational demands present …
remarkable success. However, their formidable size and computational demands present …
BiomedGPT: a unified and generalist biomedical generative pre-trained transformer for vision, language, and multimodal tasks
In this paper, we introduce a unified and generalist Biomedical Generative Pre-trained
Transformer (BiomedGPT) model, which leverages self-supervision on large and diverse …
Transformer (BiomedGPT) model, which leverages self-supervision on large and diverse …
Promptner: Prompting for named entity recognition
In a surprising turn, Large Language Models (LLMs) together with a growing arsenal of
prompt-based heuristics now offer powerful off-the-shelf approaches providing few-shot …
prompt-based heuristics now offer powerful off-the-shelf approaches providing few-shot …
3dsam-adapter: Holistic adaptation of sam from 2d to 3d for promptable medical image segmentation
Despite that the segment anything model (SAM) achieved impressive results on general-
purpose semantic segmentation with strong generalization ability on daily images, its …
purpose semantic segmentation with strong generalization ability on daily images, its …
Fine-tuning protein language models boosts predictions across diverse tasks
Prediction methods inputting embeddings from protein language models have reached or
even surpassed state-of-the-art performance on many protein prediction tasks. In natural …
even surpassed state-of-the-art performance on many protein prediction tasks. In natural …