Adapting large language models for document-level machine translation

M Wu, TT Vu, L Qu, G Foster, G Haffari - arXiv preprint arXiv:2401.06468, 2024 - arxiv.org
Large language models (LLMs) have made significant strides in various natural language
processing (NLP) tasks. Recent research shows that the moderately-sized LLMs often …

Do context-aware translation models pay the right attention?

K Yin, P Fernandes, D Pruthi, A Chaudhary… - arXiv preprint arXiv …, 2021 - arxiv.org
Context-aware machine translation models are designed to leverage contextual information,
but often fail to do so. As a result, they inaccurately disambiguate pronouns and polysemous …

Document-level language models for machine translation

F Petrick, C Herold, P Petrushkov, S Khadivi… - arXiv preprint arXiv …, 2023 - arxiv.org
Despite the known limitations, most machine translation systems today still operate on the
sentence-level. One reason for this is, that most parallel training data is only sentence-level …

Improving long context document-level machine translation

C Herold, H Ney - arXiv preprint arXiv:2306.05183, 2023 - arxiv.org
Document-level context for neural machine translation (NMT) is crucial to improve the
translation consistency and cohesion, the translation of ambiguous inputs, as well as several …

A baseline revisited: Pushing the limits of multi-segment models for context-aware translation

S Majumder, S Lauly, M Nadejde, M Federico… - arXiv preprint arXiv …, 2022 - arxiv.org
This paper addresses the task of contextual translation using multi-segment models.
Specifically we show that increasing model capacity further pushes the limits of this …

Beyond sentence-level end-to-end speech translation: Context helps

B Zhang, I Titov, B Haddow… - Proceedings of the 59th …, 2021 - aclanthology.org
Document-level contextual information has shown benefits to text-based machine
translation, but whether and how context helps end-to-end (E2E) speech translation (ST) is …

Divide and rule: Effective pre-training for context-aware multi-encoder translation models

L Lupo, M Dinarelli, L Besacier - arXiv preprint arXiv:2103.17151, 2021 - arxiv.org
Multi-encoder models are a broad family of context-aware neural machine translation
systems that aim to improve translation quality by encoding document-level contextual …

Multilingual document-level translation enables zero-shot transfer from sentences to documents

B Zhang, A Bapna, M Johnson… - arXiv preprint arXiv …, 2021 - arxiv.org
Document-level neural machine translation (DocNMT) achieves coherent translations by
incorporating cross-sentence context. However, for most language pairs there's a shortage …

Encoding sentence position in context-aware neural machine translation with concatenation

L Lupo, M Dinarelli, L Besacier - arXiv preprint arXiv:2302.06459, 2023 - arxiv.org
Context-aware translation can be achieved by processing a concatenation of consecutive
sentences with the standard Transformer architecture. This paper investigates the intuitive …

Contrastive learning for context-aware neural machine translationusing coreference information

Y Hwang, H Yun, K Jung - arXiv preprint arXiv:2109.05712, 2021 - arxiv.org
Context-aware neural machine translation (NMT) incorporates contextual information of
surrounding texts, that can improve the translation quality of document-level machine …