Parameter-efficient fine-tuning without introducing new latency
Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently
demonstrated remarkable achievements, effectively matching the performance of full fine …
demonstrated remarkable achievements, effectively matching the performance of full fine …
Ask Language Model to Clean Your Noisy Translation Data
Transformer models have demonstrated remarkable performance in neural machine
translation (NMT). However, their vulnerability to noisy input poses a significant challenge in …
translation (NMT). However, their vulnerability to noisy input poses a significant challenge in …
Is Encoder-Decoder Redundant for Neural Machine Translation?
Encoder-decoder architecture is widely adopted for sequence-to-sequence modeling tasks.
For machine translation, despite the evolution from long short-term memory networks to …
For machine translation, despite the evolution from long short-term memory networks to …
Multi-agent mutual learning at sentence-level and token-level for neural machine translation
Mutual learning, where multiple agents learn collaboratively and teach one another, has
been shown to be an effective way to distill knowledge for image classification tasks. In this …
been shown to be an effective way to distill knowledge for image classification tasks. In this …
[PDF][PDF] Language modeling and machine translation: improvements in training and modeling
G Yingbo - www-i6.informatik.rwth-aachen.de
The field of statistical language modeling and machine translation has seen rapid
developments in recent years, with artificial neural networks taking center of the stage …
developments in recent years, with artificial neural networks taking center of the stage …