Modeling latent sentence structure in neural machine translation

J Bastings, W Aziz, I Titov, K Sima'an - arXiv preprint arXiv:1901.06436, 2019 - arxiv.org
arXiv preprint arXiv:1901.06436, 2019arxiv.org
Recently it was shown that linguistic structure predicted by a supervised parser can be
beneficial for neural machine translation (NMT). In this work we investigate a more
challenging setup: we incorporate sentence structure as a latent variable in a standard NMT
encoder-decoder and induce it in such a way as to benefit the translation task. We consider
German-English and Japanese-English translation benchmarks and observe that when
using RNN encoders the model makes no or very limited use of the structure induction …
Recently it was shown that linguistic structure predicted by a supervised parser can be beneficial for neural machine translation (NMT). In this work we investigate a more challenging setup: we incorporate sentence structure as a latent variable in a standard NMT encoder-decoder and induce it in such a way as to benefit the translation task. We consider German-English and Japanese-English translation benchmarks and observe that when using RNN encoders the model makes no or very limited use of the structure induction apparatus. In contrast, CNN and word-embedding-based encoders rely on latent graphs and force them to encode useful, potentially long-distance, dependencies.
arxiv.org
以上显示的是最相近的搜索结果。 查看全部搜索结果