Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.21
|View full text |Cite
|
Sign up to set email alerts
|

New Intent Discovery with Pre-training and Contrastive Learning

Abstract: New intent discovery aims to uncover novel intent categories from user utterances to expand the set of supported intent classes. It is a critical task for the development and service expansion of a practical dialogue system. Despite its importance, this problem remains underexplored in the literature. Existing approaches typically rely on a large amount of labeled utterances and employ pseudo-labeling methods for representation learning and clustering, which are label-intensive, inefficient, and inaccurate. In… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 27 publications
0
7
0
Order By: Relevance
“…15 to balance different losses. Moreover, we adopt random token replacement (Zhang et al 2022) as data augmentation for second view generation. For other general hyperparameters, the learning rate is set to 5e −5 , training epochs are set to 80, and the batch size of labeled and unlabeled instances is set to 128 for all datasets equally.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…15 to balance different losses. Moreover, we adopt random token replacement (Zhang et al 2022) as data augmentation for second view generation. For other general hyperparameters, the learning rate is set to 5e −5 , training epochs are set to 80, and the batch size of labeled and unlabeled instances is set to 128 for all datasets equally.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…After pretraining, most existing works (Lin, Xu, and Zhang 2020;Zhang et al 2021bZhang et al , 2022 For labeled data, we take average of all instance embeddings belonging to the same category as labeled prototypes…”
Section: Learning Category Prototypesmentioning
confidence: 99%
“…is the number of total categories. We presume prior knowledge of K following previous works (Zhang et al 2021b(Zhang et al , 2022 to make a fair comparison and we tackle the problem of estimating this parameter in the experiment. Then we take average of all instance embeddings belonging to the same cluster as unlabeled prototypes…”
Section: Learning Category Prototypesmentioning
confidence: 99%
See 1 more Smart Citation
“…Thus, we propose a KCL loss to keep large interclass variance and help downstream transfer. For OOD clustering, Zhang et al ( , 2022 use kmeans to learn cluster assignments but ignore joint learning intent representations. Mou et al (2022) uses contrastive clustering where the instance-level contrastive loss for learning intent features has a gap with the cluster-level loss for clustering.…”
Section: Related Workmentioning
confidence: 99%