2018
DOI: 10.48550/arxiv.1804.02063
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Few-Shot Text Classification with Pre-Trained Word Embeddings and a Human in the Loop

Abstract: Most of the literature around text classification treats it as a supervised learning problem: given a corpus of labeled documents, train a classifier such that it can accurately predict the classes of unseen documents. In industry, however, it is not uncommon for a business to have entire corpora of documents where few or none have been classified, or where existing classifications have become meaningless. With web content, for example, poor taxonomy management can result in labels being applied indiscriminate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 3 publications
0
6
0
Order By: Relevance
“…Few-shot learning approach. We benchmark MFeEmb against prior work on conflict prediction, other embedding choices, and FsText, a few-shot model proposed by Bailey and Chopra (2018). Experiments were performed using a 300-dimensional version of MFeEmb where the length of all the three embeddings is the same, i.e., 100.…”
Section: Methodsmentioning
confidence: 99%
“…Few-shot learning approach. We benchmark MFeEmb against prior work on conflict prediction, other embedding choices, and FsText, a few-shot model proposed by Bailey and Chopra (2018). Experiments were performed using a 300-dimensional version of MFeEmb where the length of all the three embeddings is the same, i.e., 100.…”
Section: Methodsmentioning
confidence: 99%
“…Fewshot text classification entails performing classification after training or tuning a model on only a few examples. Several studies (Yu et al, 2018;Bailey and Chopra, 2018;Geng et al, 2020) have explored various approaches for few-shot text classification, which mainly involve the traditional machine learning techniques for selecting the optimal category sub-samples.…”
Section: Few-shot Text Classificationmentioning
confidence: 99%
“…Word embeddings are a powerful tool and are applied in variety of Natural Language Processing tasks, such as text classification (Aydogan and Karci, 2020;Alwehaibi and Roy, 2018;Jo and Cinarel, 2019;Bailey and Chopra, 2018;Rescigno et al, 2020) and sentiment analysis (Araque et al, 2017;Rezaeinia et al, 2019;Fu et al, 2017;Ren et al, 2016;Tang et al, 2014). However, analogies such as "Man is to computer programmer as woman is to homemaker" (Bolukbasi et al, 2016a) contain worrisome biases that are present in society and hence embedded in language.…”
Section: Introductionmentioning
confidence: 99%