Neural machine translation: A review

F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …

Neural natural language generation: A survey on multilinguality, multimodality, controllability and learning

E Erdem, M Kuyu, S Yagcioglu, A Frank… - Journal of Artificial …, 2022 - jair.org
Developing artificial learning systems that can understand and generate natural language
has been one of the long-standing goals of artificial intelligence. Recent decades have …

Compositional generalization by factorizing alignment and translation

J Russin, J Jo, R O'Reilly, Y Bengio - Proceedings of the 58th …, 2020 - aclanthology.org
Standard methods in deep learning for natural language processing fail to capture the
compositional structure of human language that allows for systematic generalization outside …

Neural Machine Translation: A Review and Survey

F Stahlberg - arXiv preprint arXiv:1912.02047, 2019 - arxiv.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …

Findings of the second workshop on neural machine translation and generation

A Birch, A Finch, MT Luong, G Neubig… - arXiv preprint arXiv …, 2018 - arxiv.org
This document describes the findings of the Second Workshop on Neural Machine
Translation and Generation, held in concert with the annual conference of the Association …

Inducing constituency trees through neural machine translation

PM Htut, K Cho, SR Bowman - arXiv preprint arXiv:1909.10056, 2019 - arxiv.org
Latent tree learning (LTL) methods learn to parse sentences using only indirect supervision
from a downstream task. Recent advances in latent tree learning have made it possible to …

The roles of language models and hierarchical models in neural sequence-to-sequence prediction

F Stahlberg - 2020 - repository.cam.ac.uk
With the advent of deep learning, research in many areas of machine learning is converging
towards the same set of methods and models. For example, long short-term memory …

Structurally comparative hinge loss for dependency-based neural text representation

K Wang, Y Zhou, J Zhang, S Wang… - ACM Transactions on …, 2020 - dl.acm.org
Dependency-based graph convolutional networks (DepGCNs) are proven helpful for text
representation to handle many natural language tasks. Almost all previous models are …