Analysis of community question‐answering issues via machine learning and deep learning: State‐of‐the‐art review
Over the last couple of decades, community question‐answering sites (CQAs) have been a
topic of much academic interest. Scholars have often leveraged traditional machine learning …
topic of much academic interest. Scholars have often leveraged traditional machine learning …
MTEB: Massive text embedding benchmark
Text embeddings are commonly evaluated on a small set of datasets from a single task not
covering their possible applications to other tasks. It is unclear whether state-of-the-art …
covering their possible applications to other tasks. It is unclear whether state-of-the-art …
A holistic approach to undesired content detection in the real world
We present a holistic approach to building a robust and useful natural language
classification system for real-world content moderation. The success of such a system relies …
classification system for real-world content moderation. The success of such a system relies …
Neural unsupervised domain adaptation in NLP---a survey
Deep neural networks excel at learning from labeled data and achieve state-of-the-art
resultson a wide array of Natural Language Processing tasks. In contrast, learning from …
resultson a wide array of Natural Language Processing tasks. In contrast, learning from …
Augmented SBERT: Data augmentation method for improving bi-encoders for pairwise sentence scoring tasks
There are two approaches for pairwise sentence scoring: Cross-encoders, which perform full-
attention over the input pair, and Bi-encoders, which map each input independently to a …
attention over the input pair, and Bi-encoders, which map each input independently to a …
Interact before align: Leveraging cross-modal knowledge for domain adaptive action recognition
Unsupervised domain adaptive video action recognition aims to recognize actions of a
target domain using a model trained with only out-of-domain (source) annotations. The …
target domain using a model trained with only out-of-domain (source) annotations. The …
We need to talk about random splits
Gorman and Bedrick (2019) argued for using random splits rather than standard splits in
NLP experiments. We argue that random splits, like standard splits, lead to overly optimistic …
NLP experiments. We argue that random splits, like standard splits, lead to overly optimistic …
Quantum transfer learning for acceptability judgements
Hybrid quantum-classical classifiers promise to positively impact critical aspects of natural
language processing tasks, particularly classification-related ones. Among the possibilities …
language processing tasks, particularly classification-related ones. Among the possibilities …
UDALM: Unsupervised domain adaptation through language modeling
C Karouzos, G Paraskevopoulos… - arXiv preprint arXiv …, 2021 - arxiv.org
In this work we explore Unsupervised Domain Adaptation (UDA) of pretrained language
models for downstream tasks. We introduce UDALM, a fine-tuning procedure, using a mixed …
models for downstream tasks. We introduce UDALM, a fine-tuning procedure, using a mixed …
Robust zero-shot cross-domain slot filling with example values
Task-oriented dialog systems increasingly rely on deep learning-based slot filling models,
usually needing extensive labeled training data for target domains. Often, however, little to …
usually needing extensive labeled training data for target domains. Often, however, little to …