Scaling instruction-finetuned language models

HW Chung, L Hou, S Longpre, B Zoph, Y Tay… - Journal of Machine …, 2024 - jmlr.org
Finetuning language models on a collection of datasets phrased as instructions has been
shown to improve model performance and generalization to unseen tasks. In this paper we …

Lila: A unified benchmark for mathematical reasoning

S Mishra, M Finlayson, P Lu, L Tang, S Welleck… - arXiv preprint arXiv …, 2022 - arxiv.org
Mathematical reasoning skills are essential for general-purpose intelligent systems to
perform tasks from grocery shopping to climate modeling. Towards evaluating and …

In-boxbart: Get instructions into biomedical multi-task learning

M Parmar, S Mishra, M Purohit, M Luo… - arXiv preprint arXiv …, 2022 - arxiv.org
Single-task models have proven pivotal in solving specific tasks; however, they have
limitations in real-world applications where multi-tasking is necessary and domain shifts are …

Maybe only 0.5% data is needed: A preliminary exploration of low training data instruction tuning

H Chen, Y Zhang, Q Zhang, H Yang, X Hu, X Ma… - arXiv preprint arXiv …, 2023 - arxiv.org
Instruction tuning for large language models (LLMs) has gained attention from researchers
due to its ability to unlock the potential of LLMs in following instructions. While instruction …

Is a question decomposition unit all we need?

P Patel, S Mishra, M Parmar, C Baral - arXiv preprint arXiv:2205.12538, 2022 - arxiv.org
Large Language Models (LMs) have achieved state-of-the-art performance on many Natural
Language Processing (NLP) benchmarks. With the growing number of new benchmarks, we …

Instruction tuned models are quick learners

H Gupta, SA Sawant, S Mishra, M Nakamura… - arXiv preprint arXiv …, 2023 - arxiv.org
Instruction tuning of language models has demonstrated the ability to enhance model
generalization to unseen tasks via in-context learning using a few examples. However …

Instructional fingerprinting of large language models

J Xu, F Wang, MD Ma, PW Koh, C Xiao… - arXiv preprint arXiv …, 2024 - arxiv.org
The exorbitant cost of training Large language models (LLMs) from scratch makes it
essential to fingerprint the models to protect intellectual property via ownership …

Help me think: A simple prompting strategy for non-experts to create customized content with models

S Mishra, E Nouri - arXiv preprint arXiv:2208.08232, 2022 - arxiv.org
Controlling the text generated by language models and customizing the content has been a
long-standing challenge. Existing prompting techniques proposed in pursuit of providing …

Biotabqa: Instruction learning for biomedical table question answering

M Luo, S Saxena, S Mishra, M Parmar… - arXiv preprint arXiv …, 2022 - arxiv.org
Table Question Answering (TQA) is an important but under-explored task. Most of the
existing QA datasets are in unstructured text format and only few of them use tables as the …

Context-ner: Contextual phrase generation at scale

H Gupta, S Verma, S Mashetty, S Mishra - arXiv preprint arXiv:2109.08079, 2021 - arxiv.org
Named Entity Recognition (NER) has seen significant progress in recent years, with
numerous state-of-the-art (SOTA) models achieving high performance. However, very few …