Zero-shot text classification with self-training

A Gera, A Halfon, E Shnarch, Y Perlitz… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent advances in large pretrained language models have increased attention to zero-shot
text classification. In particular, models finetuned on natural language inference datasets …

Teleclass: Taxonomy enrichment and llm-enhanced hierarchical text classification with minimal supervision

Y Zhang, R Yang, X Xu, R Li, J Xiao, J Shen… - arXiv preprint arXiv …, 2024 - arxiv.org
Hierarchical text classification aims to categorize each document into a set of classes in a
label taxonomy. Most earlier works focus on fully or semi-supervised methods that require a …

Dp-ssl: Towards robust semi-supervised learning with a few labeled samples

Y Xu, J Ding, L Zhang, S Zhou - Advances in Neural …, 2021 - proceedings.neurips.cc
The scarcity of labeled data is a critical obstacle to deep learning. Semi-supervised learning
(SSL) provides a promising way to leverage unlabeled data by pseudo labels. However …

Bag-of-words vs. graph vs. sequence in text classification: Questioning the necessity of text-graphs and the surprising strength of a wide MLP

L Galke, A Scherp - arXiv preprint arXiv:2109.03777, 2021 - arxiv.org
Graph neural networks have triggered a resurgence of graph-based text classification
methods, defining today's state of the art. We show that a wide multi-layer perceptron (MLP) …

BoW-based neural networks vs. cutting-edge models for single-label text classification

HI Abdalla, AA Amer, SD Ravana - Neural Computing and Applications, 2023 - Springer
To reliably and accurately classify complicated" big" datasets, machine learning models
must be continually improved. This research proposes straightforward yet competitive neural …

Beyond prompting: Making pre-trained language models better zero-shot learners by clustering representations

Y Fei, P Nie, Z Meng, R Wattenhofer… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent work has demonstrated that pre-trained language models (PLMs) are zero-shot
learners. However, most existing zero-shot methods involve heavy human engineering or …

Weakly supervised multi-label classification of full-text scientific papers

Y Zhang, B Jin, X Chen, Y Shen, Y Zhang… - Proceedings of the 29th …, 2023 - dl.acm.org
Instead of relying on human-annotated training samples to build a classifier, weakly
supervised scientific paper classification aims to classify papers only using category …

Reproducibility in computational linguistics: Is source code enough?

M Arvan, L Pina, N Parde - … of the 2022 Conference on Empirical …, 2022 - aclanthology.org
The availability of source code has been put forward as one of the most critical factors for
improving the reproducibility of scientific research. This work studies trends in source code …

A benchmark on extremely weakly supervised text classification: Reconcile seed matching and prompting approaches

Z Wang, T Wang, D Mekala, J Shang - arXiv preprint arXiv:2305.12749, 2023 - arxiv.org
Etremely Weakly Supervised Text Classification (XWS-TC) refers to text classification based
on minimal high-level human guidance, such as a few label-indicative seed words or …

LogLG: Weakly Supervised Log Anomaly Detection via Log-Event Graph Construction

H Guo, Y Guo, J Yang, J Liu, Z Li, T Zheng… - … on Database Systems …, 2023 - Springer
Fully supervised log anomaly detection methods suffer the heavy burden of annotating
massive unlabeled log data. Recently, many semi-supervised methods have been proposed …