Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.128
|View full text |Cite
|
Sign up to set email alerts
|

Few-shot Slot Tagging with Collapsed Dependency Transfer and Label-enhanced Task-adaptive Projection Network

Abstract: In this paper, we explore the slot tagging with only a few labeled support sentences (a.k.a. few-shot). Few-shot slot tagging faces a unique challenge compared to the other fewshot classification problems as it calls for modeling the dependencies between labels. But it is hard to apply previously learned label dependencies to an unseen domain, due to the discrepancy of label sets. To tackle this, we introduce a collapsed dependency transfer mechanism into the conditional random field (CRF) to transfer abstract… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
217
1

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 139 publications
(257 citation statements)
references
References 37 publications
2
217
1
Order By: Relevance
“…Prior work (Fritzler et al, 2019;Hou et al, 2020) on few-shot NER followed few-shot classification literature and adopted the episode evaluation methodology. Specifically, a NER system is evaluated with respect to multiple evaluation episodes.…”
Section: A Standard Evaluation Setupmentioning
confidence: 99%
See 4 more Smart Citations
“…Prior work (Fritzler et al, 2019;Hou et al, 2020) on few-shot NER followed few-shot classification literature and adopted the episode evaluation methodology. Specifically, a NER system is evaluated with respect to multiple evaluation episodes.…”
Section: A Standard Evaluation Setupmentioning
confidence: 99%
“…Most meta-learning approaches (Snell et al, 2017;Hou et al, 2020) simulate the test time setup during training. Hence, these approaches sample multiple support sets and test sets from the training data and learn representations to minimize their corresponding few-shot loss on the source domain.…”
Section: Pre-trained Ner Models As Token Embeddersmentioning
confidence: 99%
See 3 more Smart Citations