Instruction tuning for large language models: A survey

S Zhang, L Dong, X Li, S Zhang, X Sun, S Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
This paper surveys research works in the quickly advancing field of instruction tuning (IT), a
crucial technique to enhance the capabilities and controllability of large language models …

Large ai models in health informatics: Applications, challenges, and the future

J Qiu, L Li, J Sun, J Peng, P Shi… - IEEE Journal of …, 2023 - ieeexplore.ieee.org
Large AI models, or foundation models, are models recently emerging with massive scales
both parameter-wise and data-wise, the magnitudes of which can reach beyond billions …

A survey of large language models

WX Zhao, K Zhou, J Li, T Tang, X Wang, Y Hou… - arXiv preprint arXiv …, 2023 - arxiv.org
Language is essentially a complex, intricate system of human expressions governed by
grammatical rules. It poses a significant challenge to develop capable AI algorithms for …

Enhancing chat language models by scaling high-quality instructional conversations

N Ding, Y Chen, B Xu, Y Qin, Z Zheng, S Hu… - arXiv preprint arXiv …, 2023 - arxiv.org
Fine-tuning on instruction data has been widely validated as an effective practice for
implementing chat language models like ChatGPT. Scaling the diversity and quality of such …

Language models are super mario: Absorbing abilities from homologous models as a free lunch

L Yu, B Yu, H Yu, F Huang, Y Li - Forty-first International Conference …, 2024 - openreview.net
In this paper, we unveil that Language Models (LMs) can acquire new capabilities by
assimilating parameters from homologous models without retraining or GPUs. We first …

A survey on model compression for large language models

X Zhu, J Li, Y Liu, C Ma, W Wang - arXiv preprint arXiv:2308.07633, 2023 - arxiv.org
Large Language Models (LLMs) have revolutionized natural language processing tasks with
remarkable success. However, their formidable size and computational demands present …

BiomedGPT: a unified and generalist biomedical generative pre-trained transformer for vision, language, and multimodal tasks

K Zhang, J Yu, Z Yan, Y Liu, E Adhikarla, S Fu… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we introduce a unified and generalist Biomedical Generative Pre-trained
Transformer (BiomedGPT) model, which leverages self-supervision on large and diverse …

Promptner: Prompting for named entity recognition

D Ashok, ZC Lipton - arXiv preprint arXiv:2305.15444, 2023 - arxiv.org
In a surprising turn, Large Language Models (LLMs) together with a growing arsenal of
prompt-based heuristics now offer powerful off-the-shelf approaches providing few-shot …

3dsam-adapter: Holistic adaptation of sam from 2d to 3d for promptable medical image segmentation

S Gong, Y Zhong, W Ma, J Li, Z Wang, J Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Despite that the segment anything model (SAM) achieved impressive results on general-
purpose semantic segmentation with strong generalization ability on daily images, its …

Fine-tuning protein language models boosts predictions across diverse tasks

R Schmirler, M Heinzinger, B Rost - Nature Communications, 2024 - nature.com
Prediction methods inputting embeddings from protein language models have reached or
even surpassed state-of-the-art performance on many protein prediction tasks. In natural …