Cross-lingual language model pretraining
Recent studies have demonstrated the efficiency of generative pretraining for English
natural language understanding. In this work, we extend this approach to multiple …
natural language understanding. In this work, we extend this approach to multiple …
Cross-lingual language model pretraining
Recent studies have demonstrated the efficiency of generative pretraining for English
natural language understanding. In this work, we extend this approach to multiple …
natural language understanding. In this work, we extend this approach to multiple …
Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond
We introduce an architecture to learn joint multilingual sentence representations for 93
languages, belonging to more than 30 different families and written in 28 different scripts …
languages, belonging to more than 30 different families and written in 28 different scripts …
The state of the art in semantic representation
O Abend, A Rappoport - Proceedings of the 55th Annual Meeting …, 2017 - aclanthology.org
Semantic representation is receiving growing attention in NLP in the past few years, and
many proposals for semantic schemes (eg, AMR, UCCA, GMB, UDS) have been put forth …
many proposals for semantic schemes (eg, AMR, UCCA, GMB, UDS) have been put forth …
XNLI: Evaluating cross-lingual sentence representations
State-of-the-art natural language processing systems rely on supervision in the form of
annotated data to learn competent models. These models are generally trained on data in a …
annotated data to learn competent models. These models are generally trained on data in a …
Word translation without parallel data
State-of-the-art methods for learning cross-lingual word embeddings have relied on
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …
[PDF][PDF] Word translation without parallel data
State-of-the-art methods for learning cross-lingual word embeddings have relied on
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …
A survey of cross-lingual word embedding models
Cross-lingual representations of words enable us to reason about word meaning in
multilingual contexts and are a key facilitator of cross-lingual transfer when developing …
multilingual contexts and are a key facilitator of cross-lingual transfer when developing …
An overview of word and sense similarity
R Navigli, F Martelli - Natural Language Engineering, 2019 - cambridge.org
Over the last two decades, determining the similarity between words as well as between
their meanings, that is, word senses, has been proven to be of vital importance in the field of …
their meanings, that is, word senses, has been proven to be of vital importance in the field of …
Offline bilingual word vectors, orthogonal transformations and the inverted softmax
Usually bilingual word vectors are trained" online". Mikolov et al. showed they can also be
found" offline", whereby two pre-trained embeddings are aligned with a linear …
found" offline", whereby two pre-trained embeddings are aligned with a linear …