Scaling instruction-finetuned language models
Finetuning language models on a collection of datasets phrased as instructions has been
shown to improve model performance and generalization to unseen tasks. In this paper we …
shown to improve model performance and generalization to unseen tasks. In this paper we …
Lila: A unified benchmark for mathematical reasoning
Mathematical reasoning skills are essential for general-purpose intelligent systems to
perform tasks from grocery shopping to climate modeling. Towards evaluating and …
perform tasks from grocery shopping to climate modeling. Towards evaluating and …
In-boxbart: Get instructions into biomedical multi-task learning
Single-task models have proven pivotal in solving specific tasks; however, they have
limitations in real-world applications where multi-tasking is necessary and domain shifts are …
limitations in real-world applications where multi-tasking is necessary and domain shifts are …
Maybe only 0.5% data is needed: A preliminary exploration of low training data instruction tuning
Instruction tuning for large language models (LLMs) has gained attention from researchers
due to its ability to unlock the potential of LLMs in following instructions. While instruction …
due to its ability to unlock the potential of LLMs in following instructions. While instruction …
Is a question decomposition unit all we need?
Large Language Models (LMs) have achieved state-of-the-art performance on many Natural
Language Processing (NLP) benchmarks. With the growing number of new benchmarks, we …
Language Processing (NLP) benchmarks. With the growing number of new benchmarks, we …
Instruction tuned models are quick learners
Instruction tuning of language models has demonstrated the ability to enhance model
generalization to unseen tasks via in-context learning using a few examples. However …
generalization to unseen tasks via in-context learning using a few examples. However …
Instructional fingerprinting of large language models
The exorbitant cost of training Large language models (LLMs) from scratch makes it
essential to fingerprint the models to protect intellectual property via ownership …
essential to fingerprint the models to protect intellectual property via ownership …
Help me think: A simple prompting strategy for non-experts to create customized content with models
Controlling the text generated by language models and customizing the content has been a
long-standing challenge. Existing prompting techniques proposed in pursuit of providing …
long-standing challenge. Existing prompting techniques proposed in pursuit of providing …
Biotabqa: Instruction learning for biomedical table question answering
Table Question Answering (TQA) is an important but under-explored task. Most of the
existing QA datasets are in unstructured text format and only few of them use tables as the …
existing QA datasets are in unstructured text format and only few of them use tables as the …
Context-ner: Contextual phrase generation at scale
Named Entity Recognition (NER) has seen significant progress in recent years, with
numerous state-of-the-art (SOTA) models achieving high performance. However, very few …
numerous state-of-the-art (SOTA) models achieving high performance. However, very few …