Learning to classify open intent via soft labeling and manifold mixup

Z Cheng, Z Jiang, Y Yin, C Wang… - IEEE/ACM Transactions …, 2022 - ieeexplore.ieee.org
IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2022ieeexplore.ieee.org
Open intent classification is a practical yet challenging task in dialogue systems. Its objective
is to accurately classify samples of known intents while at the same time detecting those of
open (unknown) intents. Existing methods usually use outlier detection algorithms combined
with K-class classifier to detect open intents, where K represents the class number of known
intents. Different from them, in this paper, we consider another way without using outlier
detection algorithms. Specifically, we directly train a (K+ 1)-class classifier for open intent …
Open intent classification is a practical yet challenging task in dialogue systems. Its objective is to accurately classify samples of known intents while at the same time detecting those of open (unknown) intents. Existing methods usually use outlier detection algorithms combined with K-class classifier to detect open intents, where K represents the class number of known intents. Different from them, in this paper, we consider another way without using outlier detection algorithms. Specifically, we directly train a (K+1)-class classifier for open intent classification, where the (K+1)-th class represents open intents. To address the challenge that training a (K+1)-class classifier with training samples of only K classes, we propose a deep model based on Soft Labeling and Manifold Mixup (SLMM). In our method, soft labeling is used to reshape the label distribution of the known intent samples, aiming at reducing model’s overconfident on known intents. Manifold mixup is used to generate pseudo samples for open intents, aiming at well optimizing the decision boundary of open intents. Experiments on four benchmark datasets demonstrate that our method outperforms previous methods and achieves state-of-the-art performance. All the code and data of this work can be obtained at. 1
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果