Parameter-efficient fine-tuning without introducing new latency

B Liao, Y Meng, C Monz - arXiv preprint arXiv:2305.16742, 2023 - arxiv.org
Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently
demonstrated remarkable achievements, effectively matching the performance of full fine …

Ask Language Model to Clean Your Noisy Translation Data

Q Bolding, B Liao, BJ Denis, J Luo, C Monz - arXiv preprint arXiv …, 2023 - arxiv.org
Transformer models have demonstrated remarkable performance in neural machine
translation (NMT). However, their vulnerability to noisy input poses a significant challenge in …

Is Encoder-Decoder Redundant for Neural Machine Translation?

Y Gao, C Herold, Z Yang, H Ney - arXiv preprint arXiv:2210.11807, 2022 - arxiv.org
Encoder-decoder architecture is widely adopted for sequence-to-sequence modeling tasks.
For machine translation, despite the evolution from long short-term memory networks to …

Multi-agent mutual learning at sentence-level and token-level for neural machine translation

B Liao, Y Gao, H Ney - Findings of the Association for …, 2020 - aclanthology.org
Mutual learning, where multiple agents learn collaboratively and teach one another, has
been shown to be an effective way to distill knowledge for image classification tasks. In this …

[PDF][PDF] Language modeling and machine translation: improvements in training and modeling

G Yingbo - www-i6.informatik.rwth-aachen.de
The field of statistical language modeling and machine translation has seen rapid
developments in recent years, with artificial neural networks taking center of the stage …