[HTML][HTML] Transforming machine translation: a deep learning system reaches news translation quality comparable to human professionals

M Popel, M Tomkova, J Tomek, Ł Kaiser… - Nature …, 2020 - nature.com
The quality of human translation was long thought to be unattainable for computer
translation systems. In this study, we present a deep-learning system, CUBBITT, which …

Reformulating unsupervised style transfer as paraphrase generation

K Krishna, J Wieting, M Iyyer - arXiv preprint arXiv:2010.05700, 2020 - arxiv.org
Modern NLP defines the task of style transfer as modifying the style of a given sentence
without appreciably changing its semantics, which implies that the outputs of style transfer …

Adversarial example generation with syntactically controlled paraphrase networks

M Iyyer, J Wieting, K Gimpel, L Zettlemoyer - arXiv preprint arXiv …, 2018 - arxiv.org
We propose syntactically controlled paraphrase networks (SCPNs) and use them to
generate adversarial examples. Given a sentence and a target syntactic form (eg, a …

Findings of the 2017 conference on machine translation (wmt17)

O Bojar, R Chatterjee, C Federmann, Y Graham… - 2017 - doras.dcu.ie
This paper presents the results of the WMT17 shared tasks, which included three machine
translation (MT) tasks (news, biomedical, and multimodal), two evaluation tasks (metrics and …

ParaCrawl: Web-scale acquisition of parallel corpora

M Bañón, P Chen, B Haddow, K Heafield, H Hoang… - 2020 - strathprints.strath.ac.uk
We report on methods to create the largest publicly available parallel corpora by crawling
the web, using open source software. We empirically compare alternative methods and …

Findings of the 2016 conference on machine translation (wmt16)

O Bojar, R Chatterjee, C Federmann… - First conference on …, 2016 - research.ed.ac.uk
This paper presents the results of the WMT16 shared tasks, which included five machine
translation (MT) tasks (standard news, IT-domain, biomedical, multimodal, pronoun), three …

Training tips for the transformer model

M Popel, O Bojar - arXiv preprint arXiv:1804.00247, 2018 - arxiv.org
This article describes our experiments in neural machine translation using the recent
Tensor2Tensor framework and the Transformer sequence-to-sequence model (Vaswani et …

ParaNMT-50M: Pushing the limits of paraphrastic sentence embeddings with millions of machine translations

J Wieting, K Gimpel - arXiv preprint arXiv:1711.05732, 2017 - arxiv.org
We describe PARANMT-50M, a dataset of more than 50 million English-English sentential
paraphrase pairs. We generated the pairs automatically by using neural machine translation …

Edinburgh neural machine translation systems for WMT 16

R Sennrich, B Haddow, A Birch - arXiv preprint arXiv:1606.02891, 2016 - arxiv.org
We participated in the WMT 2016 shared news translation task by building neural translation
systems for four language pairs, each trained in both directions: English<-> Czech, English< …

Trivial transfer learning for low-resource neural machine translation

T Kocmi, O Bojar - arXiv preprint arXiv:1809.00357, 2018 - arxiv.org
Transfer learning has been proven as an effective technique for neural machine translation
under low-resource conditions. Existing methods require a common target language …