Continual lifelong learning in natural language processing: A survey

M Biesialska, K Biesialska, MR Costa-Jussa - arXiv preprint arXiv …, 2020 - arxiv.org
Continual learning (CL) aims to enable information systems to learn from a continuous data
stream across time. However, it is difficult for existing deep learning architectures to learn a …

A survey on document-level neural machine translation: Methods and evaluation

S Maruf, F Saleh, G Haffari - ACM Computing Surveys (CSUR), 2021 - dl.acm.org
Machine translation (MT) is an important task in natural language processing (NLP), as it
automates the translation process and reduces the reliance on human translators. With the …

Document-level machine translation with large language models

L Wang, C Lyu, T Ji, Z Zhang, D Yu, S Shi… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) such as ChatGPT can produce coherent, cohesive, relevant,
and fluent answers for various natural language processing (NLP) tasks. Taking document …

[PDF][PDF] Multilingual denoising pre-training for neural machine translation

Y Liu - arXiv preprint arXiv:2001.08210, 2020 - fq.pkwyx.com
This paper demonstrates that multilingual denoising pre-training produces significant
performance gains across a wide variety of machine translation (MT) tasks. We present …

Document-level neural machine translation with hierarchical attention networks

L Miculicich, D Ram, N Pappas… - arXiv preprint arXiv …, 2018 - arxiv.org
Neural Machine Translation (NMT) can be improved by including document-level contextual
information. For this purpose, we propose a hierarchical attention model to capture the …

Improving the transformer translation model with document-level context

J Zhang, H Luan, M Sun, F Zhai, J Xu, M Zhang… - arXiv preprint arXiv …, 2018 - arxiv.org
Although the Transformer translation model (Vaswani et al., 2017) has achieved state-of-the-
art performance in a variety of translation tasks, how to use document-level context to deal …

Selective attention for context-aware neural machine translation

S Maruf, AFT Martins, G Haffari - arXiv preprint arXiv:1903.08788, 2019 - arxiv.org
Despite the progress made in sentence-level NMT, current systems still fall short at
achieving fluent, good quality translation for a full document. Recent works in context-aware …

BLEU might be guilty but references are not innocent

M Freitag, D Grangier, I Caswell - arXiv preprint arXiv:2004.06063, 2020 - arxiv.org
The quality of automatic metrics for machine translation has been increasingly called into
question, especially for high-quality systems. This paper demonstrates that, while choice of …

Integrating transformer and paraphrase rules for sentence simplification

S Zhao, R Meng, D He, S Andi, P Bambang - arXiv preprint arXiv …, 2018 - arxiv.org
Sentence simplification aims to reduce the complexity of a sentence while retaining its
original meaning. Current models for sentence simplification adopted ideas from ma-chine …

Transformer: A general framework from machine translation to others

Y Zhao, J Zhang, C Zong - Machine Intelligence Research, 2023 - Springer
Abstract Machine translation is an important and challenging task that aims at automatically
translating natural language sentences from one language into another. Recently …