Recent advances in natural language processing via large pre-trained language models: A survey

B Min, H Ross, E Sulem, APB Veyseh… - ACM Computing …, 2023 - dl.acm.org
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …

Machine knowledge: Creation and curation of comprehensive knowledge bases

G Weikum, XL Dong, S Razniewski… - … and Trends® in …, 2021 - nowpublishers.com
Equipping machines with comprehensive knowledge of the world's entities and their
relationships has been a longstanding goal of AI. Over the last decade, large-scale …

Intermediate-task transfer learning with pretrained models for natural language understanding: When and why does it work?

Y Pruksachatkun, J Phang, H Liu, PM Htut… - arXiv preprint arXiv …, 2020 - arxiv.org
While pretrained models such as BERT have shown large gains across natural language
understanding tasks, their performance can be improved by further training the model on a …

MRQA 2019 shared task: Evaluating generalization in reading comprehension

A Fisch, A Talmor, R Jia, M Seo, E Choi… - arXiv preprint arXiv …, 2019 - arxiv.org
We present the results of the Machine Reading for Question Answering (MRQA) 2019
shared task on evaluating the generalization capabilities of reading comprehension …

Supervised open information extraction

G Stanovsky, J Michael, L Zettlemoyer… - Proceedings of the …, 2018 - aclanthology.org
We present data and methods that enable a supervised learning approach to Open
Information Extraction (Open IE). Central to the approach is a novel formulation of Open IE …

AmbigQA: Answering ambiguous open-domain questions

S Min, J Michael, H Hajishirzi, L Zettlemoyer - arXiv preprint arXiv …, 2020 - arxiv.org
Ambiguity is inherent to open-domain question answering; especially when exploring new
topics, it can be difficult to ask questions that have a single, unambiguous answer. In this …

Break It Down: A Question Understanding Benchmark

T Wolfson, M Geva, A Gupta, M Gardner… - Transactions of the …, 2020 - direct.mit.edu
Understanding natural language questions entails the ability to break down a question into
the requisite steps for computing its answer. In this work, we introduce a Question …

Evaluating factuality in generation with dependency-level entailment

T Goyal, G Durrett - arXiv preprint arXiv:2010.05478, 2020 - arxiv.org
Despite significant progress in text generation models, a serious limitation is their tendency
to produce text that is factually inconsistent with information in the input. Recent work has …

Zero-shot event extraction via transfer learning: Challenges and insights

Q Lyu, H Zhang, E Sulem, D Roth - … of the 59th Annual Meeting of …, 2021 - aclanthology.org
Event extraction has long been a challenging task, addressed mostly with supervised
methods that require expensive annotation and are not extensible to new event ontologies …

Transforming question answering datasets into natural language inference datasets

D Demszky, K Guu, P Liang - arXiv preprint arXiv:1809.02922, 2018 - arxiv.org
Existing datasets for natural language inference (NLI) have propelled research on language
understanding. We propose a new method for automatically deriving NLI datasets from the …