2021
DOI: 10.48550/arxiv.2107.03451
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Anticipating Safety Issues in E2E Conversational AI: Framework and Tooling

Abstract: Warning: this paper contains example data that may be offensive or upsetting.Over the last several years, end-to-end neural conversational agents have vastly improved in their ability to carry a chit-chat conversation with humans. However, these models are often trained on large datasets from the internet, and as a result, may learn undesirable behaviors from this data, such as toxic or otherwise harmful language. Researchers must thus wrestle with the issue of how and when to release these models. In this pap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(26 citation statements)
references
References 98 publications
0
23
0
Order By: Relevance
“…Similar issues have also been discussed specifically for dialog models [53]. For instance, examples of bias, offensiveness, and hate speech have been found both in training data drawn from social media, and consequently in the output of dialog models trained on such data [83].…”
Section: Related Workmentioning
confidence: 74%
See 1 more Smart Citation
“…Similar issues have also been discussed specifically for dialog models [53]. For instance, examples of bias, offensiveness, and hate speech have been found both in training data drawn from social media, and consequently in the output of dialog models trained on such data [83].…”
Section: Related Workmentioning
confidence: 74%
“…Safety and safety of dialog models: Inappropriate and unsafe risks and behaviors of language models have been extensively discussed and studied in previous works (e.g., [53,54]). Issues encountered include toxicity (e.g., [55,56,57]), bias (e.g., [58,59,60,61,62,63,64,65,66,67,68,69,70,71,72]), and inappropriately revealing personally identifying information (PII) from training data [73].…”
Section: Related Workmentioning
confidence: 99%
“…Figure 7 (Left) shows that in recent years the compute required for large-scale AI experiments has increased by more than 300, 000X relative to a decade ago. 19 Along with this rise in resource intensity, we see a corresponding (and sharp) fall in the proportion of these results that come from academia (Figure 7, Right). This suggests that, although academics may be strongly motivated by scientific curiosity, and well-poised to research safety issues, they may be significantly challenged by the high financial and engineering costs.…”
Section: Rising Gap Between Industry and Academiamentioning
confidence: 80%
“…This lack of standards compounds the problems caused by the four distinguishing features of generative models we identify in Section 2, as well as the safety issues discussed above. At the same time, there's a growing field of research oriented around identifying the weaknesses of these models, as well as potential problems with their associated development practices [7,67,9,19,72,41,50,62,66].…”
Section: Lack Of Standards and Normsmentioning
confidence: 99%
See 1 more Smart Citation