2021
DOI: 10.1609/aaai.v35i15.17612
|View full text |Cite
|
Sign up to set email alerts
|

Revisiting Mahalanobis Distance for Transformer-Based Out-of-Domain Detection

Abstract: Real-life applications, heavily relying on machine learning, such as dialog systems, demand for out-of-domain detection methods. Intent classification models should be equipped with a mechanism to distinguish seen intents from unseen ones so that the dialog agent is capable of rejecting the latter and avoiding undesired behavior. However, despite increasing attention paid to the task, the best practices for out-of-domain intent detection have not yet been fully established. This paper conducts a thorough com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 18 publications
(17 citation statements)
references
References 19 publications
0
17
0
Order By: Relevance
“…Nonetheless, almost all of the current approaches are assessed under certain assumptions about the type of OOD texts. One line of works creates in-distribution (ID) and OOD pairs from arbitrary datasets for different tasks (Hendrycks et al, 2020), while another line assumes that OOD data belong to classes in the ID task but unseen during training, e.g., in intent recognition (Podolskiy et al, 2021). Arora et al (2021) reveal the inconsistency among the evaluation protocols and category the distribution shifts to non-semantic shifts (NSS) and semantic shifts (SS), but a thorough comparison of existing methods in different settings is missing as later works either focus on either detecting NSS (Duan et al, 2022) or SS (Zhou et al, 2022).…”
Section: Far95 In Nss and Ss Scenariosmentioning
confidence: 99%
See 4 more Smart Citations
“…Nonetheless, almost all of the current approaches are assessed under certain assumptions about the type of OOD texts. One line of works creates in-distribution (ID) and OOD pairs from arbitrary datasets for different tasks (Hendrycks et al, 2020), while another line assumes that OOD data belong to classes in the ID task but unseen during training, e.g., in intent recognition (Podolskiy et al, 2021). Arora et al (2021) reveal the inconsistency among the evaluation protocols and category the distribution shifts to non-semantic shifts (NSS) and semantic shifts (SS), but a thorough comparison of existing methods in different settings is missing as later works either focus on either detecting NSS (Duan et al, 2022) or SS (Zhou et al, 2022).…”
Section: Far95 In Nss and Ss Scenariosmentioning
confidence: 99%
“…Notably, the detec-tors based on the pre-trained features, e.g., the Mahalanobis distance detector MD pre (Xu et al, 2021), excel at detecting non-semantic shifts but fail in detecting semantic shifts. In contrast, when the PLM is fine-tuned on annotated ID data, the detectors based on fine-tuned features, e.g., MD ft (Podolskiy et al, 2021), perform well in the SS scenario but disastrously fail in the NSS setting. These observations uncover an intriguing trade-off: fine-tuning contributes to the detection of semantic shifts but impairs the detection of non-semantic shifts.…”
Section: Far95 In Nss and Ss Scenariosmentioning
confidence: 99%
See 3 more Smart Citations