2021
DOI: 10.48550/arxiv.2103.12407
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Detecting Hate Speech with GPT-3

Abstract: Sophisticated language models such as OpenAI's GPT-3 can generate hateful text that targets marginalized groups. Given this capacity, we are interested in whether large language models can be used to identify hate speech and classify text as sexist or racist? We use GPT-3 to identify sexist and racist text passages with zero-, one-, and few-shot learning. We find that with zero-and one-shot learning, GPT-3 is able to identify sexist or racist text with an accuracy between 48 per cent and 69 per cent. With few-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(26 citation statements)
references
References 11 publications
(13 reference statements)
0
12
0
Order By: Relevance
“…Debouche have also raised similar concerns for the utilization of GPT-3 and recommended authors to openly share the prompts and outcomes used [5]. Other issues include embedded biases (such as hate speech towards race, sexism) [3,11], exploitation of workers for data labeling, permissions on the data used for training and environmental concerns due to the energy used in algorithm training [16].…”
Section: Gpt-3 In Academic Writingmentioning
confidence: 99%
“…Debouche have also raised similar concerns for the utilization of GPT-3 and recommended authors to openly share the prompts and outcomes used [5]. Other issues include embedded biases (such as hate speech towards race, sexism) [3,11], exploitation of workers for data labeling, permissions on the data used for training and environmental concerns due to the energy used in algorithm training [16].…”
Section: Gpt-3 In Academic Writingmentioning
confidence: 99%
“…Debouche have also raised similar concerns for the utilization of GPT-3 and recommended authors to openly share the prompts and outcomes used [5]. Other issues include embedded biases (such as hate speech towards race, sexism) [3,11], exploitation of workers for data labeling, permissions on the data used for training and environmental concerns due to the energy used in algorithm training [16].…”
Section: Gpt-3 In Academic Writingmentioning
confidence: 99%
“…Classification (6) hate speech detection [24], tweet-classifier, esrb rating, Automatically generating Request for Admissions, evaluate quiz answers, Classify news topics (AI21)…”
Section: A Identifying Llm Primitive Operations Primitive Online Demosmentioning
confidence: 99%