2022
DOI: 10.48550/arxiv.2212.10450
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Is GPT-3 a Good Data Annotator?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(20 citation statements)
references
References 0 publications
0
20
0
Order By: Relevance
“…Some of the recent research proposes text annotation or label generation using GPTs, (Generative Pretrained Transformers (GPT). The methods based on GPT have made breakthrough changes in automatic labeling tasks for data in supervised machine learning tasks [28][29][30]. With the recent launch of ChatGPT in year 2022, GPT became popular for various natural language processing (NLP) tasks.…”
Section: Limitations Of Various Topic Labeling Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Some of the recent research proposes text annotation or label generation using GPTs, (Generative Pretrained Transformers (GPT). The methods based on GPT have made breakthrough changes in automatic labeling tasks for data in supervised machine learning tasks [28][29][30]. With the recent launch of ChatGPT in year 2022, GPT became popular for various natural language processing (NLP) tasks.…”
Section: Limitations Of Various Topic Labeling Methodsmentioning
confidence: 99%
“…Limited only to customer complaint data and faces issues of data availability. [28] GPT −3-based approaches have been utilized for sequence and token level NLP tasks, and evaluation was done.…”
Section: Paper Proposed Methods Limitation [3]mentioning
confidence: 99%
“…Research exploring GLLMs for data labelling. The research community explored GLLMs for data labelling in a variety of NLP tasks like stance detection [373], [376], political tweets classification [375], sentiment analysis [376], [379], [380], hate speech detection [376], [377], bot detection [376], toxic comments detection [377], offensive comments detection [377], adverse drug reaction extraction [378], text entailment [379], topic classification [379], text generation [379], answer type classification [379], question generation [379], relation extraction [380], named entity recognition [380], [381], text summarization [382], radiology text simplification [324] etc. Most of the research works focused on English datasets, except a few research works focused on other languages like French [381], Spanish [381], Italian [381] and Basque [381].…”
Section: Data Labelling and Data Augmenta-tion Abilities Of Gllms 71 ...mentioning
confidence: 99%
“…In the past, the primary research focus was on developing specialized frameworks for specific tasks (Chiu and Nichols, 2016;Liu et al, 2016;Ding et al, 2020;Qin and Joty, 2022a). In recent years, there has been a significant shift in approach towards utilizing powerful, general-purpose language models that can be fine-tuned or prompt-tuned for a wide range of applications (Devlin et al, 2019;Yang et al, 2019;Raffel et al, 2019;Lewis et al, 2019;Brown et al, 2020;Ding et al, 2022b;Qin et al, 2023a). Through pre-training on a large-scale unlabeled corpus, pretrained language models have shown significant improvement in a wide range of NLP tasks (He et al, 2021b;Ding et al, 2022a;Qin et al, 2023b;Zhou et al, 2023).…”
Section: Retrieval-augmented Generationmentioning
confidence: 99%