Zero-shot text classification with self-training
Recent advances in large pretrained language models have increased attention to zero-shot
text classification. In particular, models finetuned on natural language inference datasets …
text classification. In particular, models finetuned on natural language inference datasets …
Teleclass: Taxonomy enrichment and llm-enhanced hierarchical text classification with minimal supervision
Hierarchical text classification aims to categorize each document into a set of classes in a
label taxonomy. Most earlier works focus on fully or semi-supervised methods that require a …
label taxonomy. Most earlier works focus on fully or semi-supervised methods that require a …
Dp-ssl: Towards robust semi-supervised learning with a few labeled samples
The scarcity of labeled data is a critical obstacle to deep learning. Semi-supervised learning
(SSL) provides a promising way to leverage unlabeled data by pseudo labels. However …
(SSL) provides a promising way to leverage unlabeled data by pseudo labels. However …
Bag-of-words vs. graph vs. sequence in text classification: Questioning the necessity of text-graphs and the surprising strength of a wide MLP
Graph neural networks have triggered a resurgence of graph-based text classification
methods, defining today's state of the art. We show that a wide multi-layer perceptron (MLP) …
methods, defining today's state of the art. We show that a wide multi-layer perceptron (MLP) …
BoW-based neural networks vs. cutting-edge models for single-label text classification
To reliably and accurately classify complicated" big" datasets, machine learning models
must be continually improved. This research proposes straightforward yet competitive neural …
must be continually improved. This research proposes straightforward yet competitive neural …
Beyond prompting: Making pre-trained language models better zero-shot learners by clustering representations
Recent work has demonstrated that pre-trained language models (PLMs) are zero-shot
learners. However, most existing zero-shot methods involve heavy human engineering or …
learners. However, most existing zero-shot methods involve heavy human engineering or …
Weakly supervised multi-label classification of full-text scientific papers
Instead of relying on human-annotated training samples to build a classifier, weakly
supervised scientific paper classification aims to classify papers only using category …
supervised scientific paper classification aims to classify papers only using category …
Reproducibility in computational linguistics: Is source code enough?
The availability of source code has been put forward as one of the most critical factors for
improving the reproducibility of scientific research. This work studies trends in source code …
improving the reproducibility of scientific research. This work studies trends in source code …
A benchmark on extremely weakly supervised text classification: Reconcile seed matching and prompting approaches
Etremely Weakly Supervised Text Classification (XWS-TC) refers to text classification based
on minimal high-level human guidance, such as a few label-indicative seed words or …
on minimal high-level human guidance, such as a few label-indicative seed words or …
LogLG: Weakly Supervised Log Anomaly Detection via Log-Event Graph Construction
Fully supervised log anomaly detection methods suffer the heavy burden of annotating
massive unlabeled log data. Recently, many semi-supervised methods have been proposed …
massive unlabeled log data. Recently, many semi-supervised methods have been proposed …