Disan: Directional self-attention network for rnn/cnn-free language understanding
Recurrent neural nets (RNN) and convolutional neural nets (CNN) are widely used on NLP
tasks to capture the long-term and local dependencies, respectively. Attention mechanisms …
tasks to capture the long-term and local dependencies, respectively. Attention mechanisms …
Skip-thought vectors
We describe an approach for unsupervised learning of a generic, distributed sentence
encoder. Using the continuity of text from books, we train an encoder-decoder model that …
encoder. Using the continuity of text from books, we train an encoder-decoder model that …
Improved semantic representations from tree-structured long short-term memory networks
Because of their superior ability to preserve sequence information over time, Long Short-
Term Memory (LSTM) networks, a type of recurrent neural network with a more complex …
Term Memory (LSTM) networks, a type of recurrent neural network with a more complex …
[PDF][PDF] Semeval-2014 task 1: Evaluation of compositional distributional semantic models on full sentences through semantic relatedness and textual entailment
This paper presents the task on the evaluation of Compositional Distributional Semantics
Models on full sentences organized for the first time within SemEval-2014. Participation was …
Models on full sentences organized for the first time within SemEval-2014. Participation was …
Siamese recurrent architectures for learning sentence similarity
J Mueller, A Thyagarajan - Proceedings of the AAAI conference on …, 2016 - ojs.aaai.org
We present a siamese adaptation of the Long Short-Term Memory (LSTM) network for
labeled data comprised of pairs of variable-length sequences. Our model is applied to …
labeled data comprised of pairs of variable-length sequences. Our model is applied to …
Word embedding for understanding natural language: a survey
Word embedding, where semantic and syntactic features are captured from unlabeled text
data, is a basic procedure in Natural Language Processing (NLP). The extracted features …
data, is a basic procedure in Natural Language Processing (NLP). The extracted features …
[PDF][PDF] Multi-perspective sentence similarity modeling with convolutional neural networks
Modeling sentence similarity is complicated by the ambiguity and variability of linguistic
expression. To cope with these challenges, we propose a model for comparing sentences …
expression. To cope with these challenges, we propose a model for comparing sentences …
[PDF][PDF] PPDB 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification
We present a new release of the Paraphrase Database. PPDB 2.0 includes a
discriminatively re-ranked set of paraphrases that achieve a higher correlation with human …
discriminatively re-ranked set of paraphrases that achieve a higher correlation with human …
[PDF][PDF] Pairwise word interaction modeling with deep neural networks for semantic similarity measurement
Textual similarity measurement is a challenging problem, as it requires understanding the
semantics of input sentences. Most previous neural network models use coarse-grained …
semantics of input sentences. Most previous neural network models use coarse-grained …
From paraphrase database to compositional paraphrase model and back
Abstract The Paraphrase Database (PPDB; Ganitkevitch et al., 2013) is an extensive
semantic resource, consisting of a list of phrase pairs with (heuristic) confidence estimates …
semantic resource, consisting of a list of phrase pairs with (heuristic) confidence estimates …