2020
DOI: 10.48550/arxiv.2011.01314
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Automatic Detection of Machine Generated Text: A Critical Survey

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(27 citation statements)
references
References 0 publications
0
21
0
Order By: Relevance
“…Fake news generated by simpler language models were also hard to detect and found to pass as human (Zellers et al, 2020). The risk of fake news generated by LMs is widely recognised and has spurred research into detecting such synthetic content (Jawahar et al, 2020). On polarisation, (McGuffie and Newhouse, 2020) demonstrated that via simple prompt engineering, GPT-3 can be used to generate content that emulates content produced by violent far-right extremist communities.…”
Section: Examplesmentioning
confidence: 99%
“…Fake news generated by simpler language models were also hard to detect and found to pass as human (Zellers et al, 2020). The risk of fake news generated by LMs is widely recognised and has spurred research into detecting such synthetic content (Jawahar et al, 2020). On polarisation, (McGuffie and Newhouse, 2020) demonstrated that via simple prompt engineering, GPT-3 can be used to generate content that emulates content produced by violent far-right extremist communities.…”
Section: Examplesmentioning
confidence: 99%
“…Moreover, (Uchendu et al, 2020) shows that RoBERTa outperforms existing detectors in detecting automatically generated news articles and product reviews which are generated by state of the art models like GPT-2. Despite the success of RoBERTa, recent research (Jawahar et al, 2020) shows that its dependence on large amounts of data, limits limits its use for detection. (Wolff and Wolff, 2020) challenges the RoBERTa model by exposing it with homoglyph and misspelling attacks and their results show a drastic drop in recall.…”
Section: Automatic Detection Of Machine Generated Textmentioning
confidence: 99%
“…Whilst it is true that existing work has been conducted to develop models capable of detecting AI-generated texts, it is worth noting that these solutions are typically limited in their generalisability. Generally, successful models are only effective in detecting AI-generated texts from a specific known model (e.g., GPT-2, XLM) of a specific type (e.g., news articles, blog posts) (Jawahar, Abdul-Mageed, and Lakshmanan 2020). There exists no "silver bullet" capable of making these potential deceptive texts trivial to identify.…”
Section: Our Contributionsmentioning
confidence: 99%
“…However, it is important to note that currently it is still possible to distinguish between AI-generated texts and human-crafted texts when one is actively looking for them, with classifiers trained to this task achieving fair-to-good performances (Jawahar, Abdul-Mageed, and Lakshmanan 2020). Given this, it is likely that AA systems in current use can mitigate the threats of NLG-based deception by being combined with some form of classifier trained specifically to the task of detecting AI-generated content.…”
Section: Broader Perspectives and Ethicsmentioning
confidence: 99%