Universal transformers

M Dehghani, S Gouws, O Vinyals, J Uszkoreit… - arXiv preprint arXiv …, 2018 - arxiv.org
Recurrent neural networks (RNNs) sequentially process data by updating their state with
each new data point, and have long been the de facto choice for sequence modeling tasks …

Reasoning with latent structure refinement for document-level relation extraction

G Nan, Z Guo, I Sekulić, W Lu - arXiv preprint arXiv:2005.06312, 2020 - arxiv.org
Document-level relation extraction requires integrating information within and across
multiple sentences of a document and capturing complex interactions between inter …

What artificial neural networks can tell us about human language acquisition

A Warstadt, SR Bowman - Algebraic structures in natural …, 2022 - taylorfrancis.com
Rapid progress in machine learning for natural language processing has the potential to
transform debates about how humans learn language. However, the learning environments …

Entity structure within and throughout: Modeling mention dependencies for document-level relation extraction

B Xu, Q Wang, Y Lyu, Y Zhu, Z Mao - … of the AAAI conference on artificial …, 2021 - ojs.aaai.org
Abstract Entities, as the essential elements in relation extraction tasks, exhibit certain
structure. In this work, we formulate such entity structure as distinctive dependencies …

Hierarchical graph network for multi-hop question answering

Y Fang, S Sun, Z Gan, R Pillai, S Wang… - arXiv preprint arXiv …, 2019 - arxiv.org
In this paper, we present Hierarchical Graph Network (HGN) for multi-hop question
answering. To aggregate clues from scattered texts across multiple paragraphs, a …

Question answering by reasoning across documents with graph convolutional networks

N De Cao, W Aziz, I Titov - arXiv preprint arXiv:1808.09920, 2018 - arxiv.org
Most research in reading comprehension has focused on answering questions based on
individual documents or even single paragraphs. We introduce a neural model which …

Multi-hop reading comprehension through question decomposition and rescoring

S Min, V Zhong, L Zettlemoyer, H Hajishirzi - arXiv preprint arXiv …, 2019 - arxiv.org
Multi-hop Reading Comprehension (RC) requires reasoning and aggregation across
several paragraphs. We propose a system for multi-hop RC that decomposes a …

Dynamically fused graph network for multi-hop reasoning

L Qiu, Y Xiao, Y Qu, H Zhou, L Li… - Proceedings of the 57th …, 2019 - aclanthology.org
Text-based question answering (TBQA) has been studied extensively in recent years. Most
existing approaches focus on finding the answer to a question within a single paragraph …

A hierarchical multi-task approach for learning embeddings from semantic tasks

V Sanh, T Wolf, S Ruder - Proceedings of the AAAI conference on …, 2019 - ojs.aaai.org
Much effort has been devoted to evaluate whether multi-task learning can be leveraged to
learn rich representations that can be used in various Natural Language Processing (NLP) …

Learning which features matter: RoBERTa acquires a preference for linguistic generalizations (eventually)

A Warstadt, Y Zhang, HS Li, H Liu… - arXiv preprint arXiv …, 2020 - arxiv.org
One reason pretraining on self-supervised linguistic tasks is effective is that it teaches
models features that are helpful for language understanding. However, we want pretrained …