Inseq: An interpretability toolkit for sequence generation models

G Sarti, N Feldhus, L Sickert, O Van Der Wal… - arXiv preprint arXiv …, 2023 - arxiv.org
Past work in natural language processing interpretability focused mainly on popular
classification tasks while largely overlooking generation settings, partly due to a lack of …

How do languages influence each other? Studying cross-lingual data sharing during LLM fine-tuning

R Choenni, D Garrette, E Shutova - arXiv preprint arXiv:2305.13286, 2023 - arxiv.org
Multilingual large language models (MLLMs) are jointly trained on data from many different
languages such that representation of individual languages can benefit from other …

Make every example count: On the stability and utility of self-influence for learning from noisy NLP datasets

I Bejan, A Sokolov, K Filippova - arXiv preprint arXiv:2302.13959, 2023 - arxiv.org
Increasingly larger datasets have become a standard ingredient to advancing the state-of-
the-art in NLP. However, data quality might have already become the bottleneck to unlock …

Examining modularity in multilingual lms via language-specialized subnetworks

R Choenni, E Shutova, D Garrette - arXiv preprint arXiv:2311.08273, 2023 - arxiv.org
Recent work has proposed explicitly inducing language-wise modularity in multilingual LMs
via sparse fine-tuning (SFT) on per-language subnetworks as a means of better guiding …

The Echoes of Multilinguality: Tracing Cultural Value Shifts during LM Fine-tuning

R Choenni, A Lauscher, E Shutova - arXiv preprint arXiv:2405.12744, 2024 - arxiv.org
Texts written in different languages reflect different culturally-dependent beliefs of their
writers. Thus, we expect multilingual LMs (MLMs), that are jointly trained on a concatenation …

The Echoes of Multilinguality: Tracing Cultural Value Shifts during Language Model Fine-tuning

R Choenni, A Lauscher, E Shutova - Proceedings of the 62nd …, 2024 - aclanthology.org
Texts written in different languages reflect different culturally-dependent beliefs of their
writers. Thus, we expect multilingual LMs (MLMs), that are jointly trained on a concatenation …