Adapting large language models for document-level machine translation
Large language models (LLMs) have made significant strides in various natural language
processing (NLP) tasks. Recent research shows that the moderately-sized LLMs often …
processing (NLP) tasks. Recent research shows that the moderately-sized LLMs often …
Do context-aware translation models pay the right attention?
Context-aware machine translation models are designed to leverage contextual information,
but often fail to do so. As a result, they inaccurately disambiguate pronouns and polysemous …
but often fail to do so. As a result, they inaccurately disambiguate pronouns and polysemous …
Document-level language models for machine translation
Despite the known limitations, most machine translation systems today still operate on the
sentence-level. One reason for this is, that most parallel training data is only sentence-level …
sentence-level. One reason for this is, that most parallel training data is only sentence-level …
Improving long context document-level machine translation
Document-level context for neural machine translation (NMT) is crucial to improve the
translation consistency and cohesion, the translation of ambiguous inputs, as well as several …
translation consistency and cohesion, the translation of ambiguous inputs, as well as several …
A baseline revisited: Pushing the limits of multi-segment models for context-aware translation
This paper addresses the task of contextual translation using multi-segment models.
Specifically we show that increasing model capacity further pushes the limits of this …
Specifically we show that increasing model capacity further pushes the limits of this …
Beyond sentence-level end-to-end speech translation: Context helps
Document-level contextual information has shown benefits to text-based machine
translation, but whether and how context helps end-to-end (E2E) speech translation (ST) is …
translation, but whether and how context helps end-to-end (E2E) speech translation (ST) is …
Divide and rule: Effective pre-training for context-aware multi-encoder translation models
Multi-encoder models are a broad family of context-aware neural machine translation
systems that aim to improve translation quality by encoding document-level contextual …
systems that aim to improve translation quality by encoding document-level contextual …
Multilingual document-level translation enables zero-shot transfer from sentences to documents
Document-level neural machine translation (DocNMT) achieves coherent translations by
incorporating cross-sentence context. However, for most language pairs there's a shortage …
incorporating cross-sentence context. However, for most language pairs there's a shortage …
Encoding sentence position in context-aware neural machine translation with concatenation
Context-aware translation can be achieved by processing a concatenation of consecutive
sentences with the standard Transformer architecture. This paper investigates the intuitive …
sentences with the standard Transformer architecture. This paper investigates the intuitive …
Contrastive learning for context-aware neural machine translationusing coreference information
Context-aware neural machine translation (NMT) incorporates contextual information of
surrounding texts, that can improve the translation quality of document-level machine …
surrounding texts, that can improve the translation quality of document-level machine …