[HTML][HTML] Neural machine translation: A review of methods, resources, and tools

Z Tan, S Wang, Z Yang, G Chen, X Huang, M Sun… - AI Open, 2020 - Elsevier
Abstract Machine translation (MT) is an important sub-field of natural language processing
that aims to translate natural languages using computers. In recent years, end-to-end neural …

Retrieval-augmented generation for knowledge-intensive nlp tasks

P Lewis, E Perez, A Piktus, F Petroni… - Advances in …, 2020 - proceedings.neurips.cc
Large pre-trained language models have been shown to store factual knowledge in their
parameters, and achieve state-of-the-art results when fine-tuned on downstream NLP tasks …

Adapting language models for zero-shot learning by meta-tuning on dataset and prompt collections

R Zhong, K Lee, Z Zhang, D Klein - arXiv preprint arXiv:2104.04670, 2021 - arxiv.org
Large pre-trained language models (LMs) such as GPT-3 have acquired a surprising ability
to perform zero-shot learning. For example, to classify sentiment without any training …

Moca: Measuring human-language model alignment on causal and moral judgment tasks

A Nie, Y Zhang, AS Amdekar, C Piech… - Advances in …, 2023 - proceedings.neurips.cc
Human commonsense understanding of the physical and social world is organized around
intuitive theories. These theories support making causal and moral judgments. When …

Robust neural machine translation with doubly adversarial inputs

Y Cheng, L Jiang, W Macherey - arXiv preprint arXiv:1906.02443, 2019 - arxiv.org
Neural machine translation (NMT) often suffers from the vulnerability to noisy perturbations
in the input. We propose an approach to improving the robustness of NMT models, which …

Faithfulness in natural language generation: A systematic survey of analysis, evaluation and optimization methods

W Li, W Wu, M Chen, J Liu, X Xiao, H Wu - arXiv preprint arXiv:2203.05227, 2022 - arxiv.org
Natural Language Generation (NLG) has made great progress in recent years due to the
development of deep learning techniques such as pre-trained language models. This …

Robust transfer learning with pretrained language models through adapters

W Han, B Pang, Y Wu - arXiv preprint arXiv:2108.02340, 2021 - arxiv.org
Transfer learning with large pretrained transformer-based language models like BERT has
become a dominating approach for most NLP tasks. Simply fine-tuning those large language …

Evaluating robustness to input perturbations for neural machine translation

X Niu, P Mathur, G Dinu, Y Al-Onaizan - arXiv preprint arXiv:2005.00580, 2020 - arxiv.org
Neural Machine Translation (NMT) models are sensitive to small perturbations in the input.
Robustness to such perturbations is typically measured using translation quality metrics …

Improving multilingual neural machine translation system for indic languages

S Bala Das, A Biradar, T Kumar Mishra… - ACM Transactions on …, 2023 - dl.acm.org
The Machine Translation System (MTS) serves as effective tool for communication by
translating text or speech from one language to another language. Recently, neural machine …

Generative adversarial neural machine translation for phonetic languages via reinforcement learning

A Kumar, A Pratap, AK Singh - IEEE Transactions on Emerging …, 2022 - ieeexplore.ieee.org
Neural Machine Translation (NMT) heavily depends on the context vectors generated via
attention network for the target word prediction. Existing works primarily focus on generating …