Chatgpt beyond english: Towards a comprehensive evaluation of large language models in multilingual learning
Over the last few years, large language models (LLMs) have emerged as the most important
breakthroughs in natural language processing (NLP) that fundamentally transform research …
breakthroughs in natural language processing (NLP) that fundamentally transform research …
XNLI: Evaluating cross-lingual sentence representations
State-of-the-art natural language processing systems rely on supervision in the form of
annotated data to learn competent models. These models are generally trained on data in a …
annotated data to learn competent models. These models are generally trained on data in a …
A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings
Recent work has managed to learn cross-lingual word embeddings without parallel data by
mapping monolingual embeddings to a shared space through adversarial training …
mapping monolingual embeddings to a shared space through adversarial training …
A survey of cross-lingual word embedding models
Cross-lingual representations of words enable us to reason about word meaning in
multilingual contexts and are a key facilitator of cross-lingual transfer when developing …
multilingual contexts and are a key facilitator of cross-lingual transfer when developing …
Learning bilingual word embeddings with (almost) no bilingual data
Most methods to learn bilingual word embeddings rely on large parallel corpora, which is
difficult to obtain for most language pairs. This has motivated an active research line to relax …
difficult to obtain for most language pairs. This has motivated an active research line to relax …
Gromov-Wasserstein alignment of word embedding spaces
D Alvarez-Melis, TS Jaakkola - arXiv preprint arXiv:1809.00013, 2018 - arxiv.org
Cross-lingual or cross-domain correspondences play key roles in tasks ranging from
machine translation to transfer learning. Recently, purely unsupervised methods operating …
machine translation to transfer learning. Recently, purely unsupervised methods operating …
Generalizing and improving bilingual word embedding mappings with a multi-step framework of linear transformations
Using a dictionary to map independently trained word embeddings to a shared space has
shown to be an effective approach to learn bilingual word embeddings. In this work, we …
shown to be an effective approach to learn bilingual word embeddings. In this work, we …
Adversarial training for unsupervised bilingual lexicon induction
Word embeddings are well known to capture linguistic regularities of the language on which
they are trained. Researchers also observe that these regularities can transfer across …
they are trained. Researchers also observe that these regularities can transfer across …
Neural cross-lingual named entity recognition with minimal resources
For languages with no annotated resources, unsupervised transfer of natural language
processing models such as named-entity recognition (NER) from resource-rich languages …
processing models such as named-entity recognition (NER) from resource-rich languages …