Neural machine translation: A review

F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …

[HTML][HTML] Neural machine translation: A review of methods, resources, and tools

Z Tan, S Wang, Z Yang, G Chen, X Huang, M Sun… - AI Open, 2020 - Elsevier
Abstract Machine translation (MT) is an important sub-field of natural language processing
that aims to translate natural languages using computers. In recent years, end-to-end neural …

Mask-predict: Parallel decoding of conditional masked language models

M Ghazvininejad, O Levy, Y Liu… - arXiv preprint arXiv …, 2019 - arxiv.org
Most machine translation systems generate text autoregressively from left to right. We,
instead, use a masked language modeling objective to train a model to predict any subset of …

A survey on non-autoregressive generation for neural machine translation and beyond

Y Xiao, L Wu, J Guo, J Li, M Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …

Glancing transformer for non-autoregressive neural machine translation

L Qian, H Zhou, Y Bao, M Wang, L Qiu… - arXiv preprint arXiv …, 2020 - arxiv.org
Recent work on non-autoregressive neural machine translation (NAT) aims at improving the
efficiency by parallel decoding without sacrificing the quality. However, existing NAT …

Flowseq: Non-autoregressive conditional sequence generation with generative flow

X Ma, C Zhou, X Li, G Neubig, E Hovy - arXiv preprint arXiv:1909.02480, 2019 - arxiv.org
Most sequence-to-sequence (seq2seq) models are autoregressive; they generate each
token by conditioning on previously generated tokens. In contrast, non-autoregressive …

Deep encoder, shallow decoder: Reevaluating non-autoregressive machine translation

J Kasai, N Pappas, H Peng, J Cross… - arXiv preprint arXiv …, 2020 - arxiv.org
Much recent effort has been invested in non-autoregressive neural machine translation,
which appears to be an efficient alternative to state-of-the-art autoregressive machine …

Fully non-autoregressive neural machine translation: Tricks of the trade

J Gu, X Kong - arXiv preprint arXiv:2012.15833, 2020 - arxiv.org
Fully non-autoregressive neural machine translation (NAT) is proposed to simultaneously
predict tokens with single forward of neural networks, which significantly reduces the …

Directed acyclic transformer for non-autoregressive machine translation

F Huang, H Zhou, Y Liu, H Li… - … Conference on Machine …, 2022 - proceedings.mlr.press
Abstract Non-autoregressive Transformers (NATs) significantly reduce the decoding latency
by generating all tokens in parallel. However, such independent predictions prevent NATs …

Non-autoregressive machine translation with latent alignments

C Saharia, W Chan, S Saxena, M Norouzi - arXiv preprint arXiv …, 2020 - arxiv.org
This paper presents two strong methods, CTC and Imputer, for non-autoregressive machine
translation that model latent alignments with dynamic programming. We revisit CTC for …