Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1581
|View full text |Cite
|
Sign up to set email alerts
|

Joint Learning of Domain Classification and Out-of-Domain Detection with Dynamic Class Weighting for Satisficing False Acceptance Rates

Abstract: In domain classification for spoken dialog systems, correct detection of out-of-domain (OOD) utterances is crucial because it reduces confusion and unnecessary interaction costs between users and the systems. Previous work usually utilizes OOD detectors that are trained separately from in-domain (IND) classifiers, and confidence thresholding for OOD detection given target evaluation scores. In this paper, we introduce a neural joint learning model for domain classification and OOD detection, where dynamic clas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
26
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 42 publications
(26 citation statements)
references
References 26 publications
0
26
0
Order By: Relevance
“…Another type of neural based OOD detection models aims to utilize a set of OOD data in the training phrase. Specifically, a special "OOD" label is added in a binary or multi-class classifier (e.g., [34]), and the inputs that fall into this special "OOD" class is rejected. However, the feasibility of this naive approach is limited in practice since suitable OOD data are usually hard to collect, and incorporating too many irrelevant OOD samples in training may cause a serious issue of data imbalance.…”
Section: Related Workmentioning
confidence: 99%
“…Another type of neural based OOD detection models aims to utilize a set of OOD data in the training phrase. Specifically, a special "OOD" label is added in a binary or multi-class classifier (e.g., [34]), and the inputs that fall into this special "OOD" class is rejected. However, the feasibility of this naive approach is limited in practice since suitable OOD data are usually hard to collect, and incorporating too many irrelevant OOD samples in training may cause a serious issue of data imbalance.…”
Section: Related Workmentioning
confidence: 99%
“…Text classification tasks in real-world applications often consists of 2 components-In-Doman (ID) classification and Out-of-Domain (OOD) detection components Kim and Kim, 2018;Shu et al, 2017;Shamekhi et al, 2018). ID classification refers to classifying a user's input with a label that exists in the training data, and OOD detection refers to designate a special OOD tag to the input when it does not belong to any of the labels in the ID training dataset (Dai et al, 2007).…”
Section: Introductionmentioning
confidence: 99%
“…For training classifiers, we initialize the embedding layer through publicly available GloVe [29] pre-trained word vectors (400 thousand words and 300 dimensions) 5 and further fine-tune the embedding layer through back-propagation. For BiLSTM model, we set the cell output dimension as 128 and dropout rate as 0.5.…”
Section: Hyperparametersmentioning
confidence: 99%
“…Only a few previous studies are related to unknown intent detection. For example, Kim and Kim [5] try to train an intent classifier and out-of-domain detector jointly, but they still need out-of-domain examples during the training process. Generative methods [6] try to generate positive and negative examples from known classes by using adversarial learning to augment training data.…”
Section: Introductionmentioning
confidence: 99%