Disan: Directional self-attention network for rnn/cnn-free language understanding

T Shen, T Zhou, G Long, J Jiang, S Pan… - Proceedings of the AAAI …, 2018 - ojs.aaai.org
Recurrent neural nets (RNN) and convolutional neural nets (CNN) are widely used on NLP
tasks to capture the long-term and local dependencies, respectively. Attention mechanisms …

Skip-thought vectors

R Kiros, Y Zhu, RR Salakhutdinov… - Advances in neural …, 2015 - proceedings.neurips.cc
We describe an approach for unsupervised learning of a generic, distributed sentence
encoder. Using the continuity of text from books, we train an encoder-decoder model that …

Improved semantic representations from tree-structured long short-term memory networks

KS Tai, R Socher, CD Manning - arXiv preprint arXiv:1503.00075, 2015 - arxiv.org
Because of their superior ability to preserve sequence information over time, Long Short-
Term Memory (LSTM) networks, a type of recurrent neural network with a more complex …

[PDF][PDF] Semeval-2014 task 1: Evaluation of compositional distributional semantic models on full sentences through semantic relatedness and textual entailment

M Marelli, L Bentivogli, M Baroni… - Proceedings of the …, 2014 - aclanthology.org
This paper presents the task on the evaluation of Compositional Distributional Semantics
Models on full sentences organized for the first time within SemEval-2014. Participation was …

Siamese recurrent architectures for learning sentence similarity

J Mueller, A Thyagarajan - Proceedings of the AAAI conference on …, 2016 - ojs.aaai.org
We present a siamese adaptation of the Long Short-Term Memory (LSTM) network for
labeled data comprised of pairs of variable-length sequences. Our model is applied to …

Word embedding for understanding natural language: a survey

Y Li, T Yang - Guide to big data applications, 2018 - Springer
Word embedding, where semantic and syntactic features are captured from unlabeled text
data, is a basic procedure in Natural Language Processing (NLP). The extracted features …

[PDF][PDF] Multi-perspective sentence similarity modeling with convolutional neural networks

H He, K Gimpel, J Lin - Proceedings of the 2015 conference on …, 2015 - aclanthology.org
Modeling sentence similarity is complicated by the ambiguity and variability of linguistic
expression. To cope with these challenges, we propose a model for comparing sentences …

[PDF][PDF] PPDB 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification

E Pavlick, P Rastogi, J Ganitkevitch… - Proceedings of the …, 2015 - aclanthology.org
We present a new release of the Paraphrase Database. PPDB 2.0 includes a
discriminatively re-ranked set of paraphrases that achieve a higher correlation with human …

[PDF][PDF] Pairwise word interaction modeling with deep neural networks for semantic similarity measurement

H He, J Lin - Proceedings of the 2016 conference of the north …, 2016 - aclanthology.org
Textual similarity measurement is a challenging problem, as it requires understanding the
semantics of input sentences. Most previous neural network models use coarse-grained …

From paraphrase database to compositional paraphrase model and back

J Wieting, M Bansal, K Gimpel… - Transactions of the …, 2015 - direct.mit.edu
Abstract The Paraphrase Database (PPDB; Ganitkevitch et al., 2013) is an extensive
semantic resource, consisting of a list of phrase pairs with (heuristic) confidence estimates …