Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.285
|View full text |Cite
|
Sign up to set email alerts
|

Task-Aware Representation of Sentences for Generic Text Classification

Abstract: State-of-the-art approaches for text classification leverage a transformer architecture with a linear layer on top that outputs a class distribution for a given prediction problem. While effective, this approach suffers from conceptual limitations that affect its utility in few-shot or zero-shot transfer learning scenarios. First, the number of classes to predict needs to be pre-defined. In a transfer learning setting, in which new classes are added to an already trained classifier, all information contained i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
42
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(57 citation statements)
references
References 18 publications
0
42
0
Order By: Relevance
“…We aim to design a personalization framework that is generalizable to arbitrary classification tasks without requiring modification to the model architecture. To that end, we draw inspiration from (Halder et al, 2020) and formulate the multiclass classification problem as a series of binary classification tasks:…”
Section: Universal Binary Classificationmentioning
confidence: 99%
See 2 more Smart Citations
“…We aim to design a personalization framework that is generalizable to arbitrary classification tasks without requiring modification to the model architecture. To that end, we draw inspiration from (Halder et al, 2020) and formulate the multiclass classification problem as a series of binary classification tasks:…”
Section: Universal Binary Classificationmentioning
confidence: 99%
“…We train for 50 epochs (unless noted otherwise) and report F1 scores on the test set. For hyper-parameters, we use a batch size of 16 and a learning rate of 0.02, following the standard in (Halder et al, 2020).…”
Section: Experimental Settingsmentioning
confidence: 99%
See 1 more Smart Citation
“…We only require the name of classes rather than manually constructed prompts or templates to convey label semantics. TARS (Halder et al, 2020) also leverages pre-trained language models and label semantics based on binary text classification. However, our method further strengthens generalization ability via meta-learning framework especially in cross-domain and fine-grained cases.…”
Section: Using Label Information For Text Classificationmentioning
confidence: 99%
“…In the manual labeling process, we achieve 77% agreement between two trained annotators and solve conflicts with a third annotator (the main author of this paper). To classify documents, we use a few-shot learning model based on Task-Aware Representation of Sentences (Halder et al, 2020), implemented in flairNLP (Akbik et al, 2019). We achieve 80.8% accuracy on the holdout set.…”
Section: Classification Of Document Functionmentioning
confidence: 99%