Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

A survey on non-autoregressive generation for neural machine translation and beyond

Y Xiao, L Wu, J Guo, J Li, M Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …

M3p: Learning universal representations via multitask multilingual multimodal pre-training

M Ni, H Huang, L Su, E Cui, T Bharti… - Proceedings of the …, 2021 - openaccess.thecvf.com
We present M3P, a Multitask Multilingual Multimodal Pre-trained model that combines
multilingual pre-training and multimodal pre-training into a unified framework via multitask …

Improving neural machine translation by bidirectional training

L Ding, D Wu, D Tao - arXiv preprint arXiv:2109.07780, 2021 - arxiv.org
We present a simple and effective pretraining strategy--bidirectional training (BiT) for neural
machine translation. Specifically, we bidirectionally update the model parameters at the …

CROP: zero-shot cross-lingual named entity recognition with multilingual labeled sequence translation

J Yang, S Huang, S Ma, Y Yin, L Dong, D Zhang… - arXiv preprint arXiv …, 2022 - arxiv.org
Named entity recognition (NER) suffers from the scarcity of annotated training data,
especially for low-resource languages without labeled data. Cross-lingual NER has been …

On the complementarity between pre-training and back-translation for neural machine translation

X Liu, L Wang, DF Wong, L Ding, LS Chao… - arXiv preprint arXiv …, 2021 - arxiv.org
Pre-training (PT) and back-translation (BT) are two simple and powerful methods to utilize
monolingual data for improving the model performance of neural machine translation (NMT) …

End-to-end speech translation for code switched speech

O Weller, M Sperber, T Pires, H Setiawan… - arXiv preprint arXiv …, 2022 - arxiv.org
Code switching (CS) refers to the phenomenon of interchangeably using words and phrases
from different languages. CS can pose significant accuracy challenges to NLP, due to the …

Multilingual agreement for multilingual neural machine translation

J Yang, Y Yin, S Ma, H Huang, D Zhang… - Proceedings of the …, 2021 - aclanthology.org
Although multilingual neural machine translation (MNMT) enables multiple language
translations, the training process is based on independent multilingual objectives. Most …

Cultural concept adaptation on multimodal reasoning

Z Li, Y Zhang - Proceedings of the 2023 Conference on Empirical …, 2023 - aclanthology.org
Developing cultural adaptation methods is important, which can improve the model
performance on the low-resource ones and provide more equitable opportunities for …

Continual mixed-language pre-training for extremely low-resource neural machine translation

Z Liu, GI Winata, P Fung - arXiv preprint arXiv:2105.03953, 2021 - arxiv.org
The data scarcity in low-resource languages has become a bottleneck to building robust
neural machine translation systems. Fine-tuning a multilingual pre-trained model (eg …