2022
DOI: 10.48550/arxiv.2201.07763
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When Is It Acceptable to Break the Rules? Knowledge Representation of Moral Judgement Based on Empirical Data

Abstract: One of the most remarkable things about the human moral mind is its flexibility. We can make moral judgments about cases we have never seen before. We can decide that pre-established rules should be broken. We can invent novel rules on the fly. Capturing this flexibility is one of the central challenges in developing AI systems that can interpret and produce human-like moral judgment. This paper details the results of a study of real-world decision makers who judge whether it is acceptable to break a well-esta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…That is, the moral sense gives us not only the duty to behave cooperatively, but also the duty to respect and enforce good "rules of the game" that incentivize efficient and fair behaviors. In line with this idea, empirical observations show that humans are indeed capable of assessing, on the fly, what is moral and immoral, even when their actions are rules whose effects are indirect, through incentives on other behaviors (Awad et al, 2022;Levine et al, 2018Levine et al, , 2020. We will call "moral rules of the game", or sometimes more simply "moral rules", these second-order contracts.…”
Section: Moral Contracts May Reach Their Social Value Indirectly and ...mentioning
confidence: 93%
See 2 more Smart Citations
“…That is, the moral sense gives us not only the duty to behave cooperatively, but also the duty to respect and enforce good "rules of the game" that incentivize efficient and fair behaviors. In line with this idea, empirical observations show that humans are indeed capable of assessing, on the fly, what is moral and immoral, even when their actions are rules whose effects are indirect, through incentives on other behaviors (Awad et al, 2022;Levine et al, 2018Levine et al, , 2020. We will call "moral rules of the game", or sometimes more simply "moral rules", these second-order contracts.…”
Section: Moral Contracts May Reach Their Social Value Indirectly and ...mentioning
confidence: 93%
“…Humans, including young children, are able to evaluate the costs and benefits of people's behaviors in various contexts, and this allows them both to predict others' behavior (Jara-Ettinger et al, 2015, and to infer others' costs and benefits functions from their actions, based on the assumption that they will seek to maximize their utility (Baker et al, 2009;Liu et al, 2017;Sosa et al, 2021). Furthermore, people make use of this ability to calculate, in context, what is right or wrong, including in cases they have never seen before (Awad et al, 2022;Carlson et al, 2022;Levine et al, 2020), and thereby to evaluate morally the behavior of others (Berman and Silver, 2022;Bigman and Tamir, 2016;Gerstenberg et al, 2018;Jara-ettinger et al, 2014;Kodipady et al;Kraft-Todd et al, 2021;Sosa et al, 2021), and manage their own reputation (Kleiman-weiner et al, 2017). Humans are thus demonstrably equipped with cognitive mechanisms to make sophisticated context-dependent assessments of costs and benefits, and use them in everyday life to assess the moral value of their actions and those of others.…”
Section: Morality As the Cognitive Organ Of Reciprocitymentioning
confidence: 99%
See 1 more Smart Citation
“…While the framework is general, here we instantiate a version for the context of dialogue agents built on Transformer-based Large Language Models (LLM). Our approach builds directly on recent work arguing for contractualism in AI alignment (Awad et al 2022;Jin et al 2022;. Our Contractual AI is also complementary to previous rule-based approaches (Forbes et al 2020;Solaiman and Dennison 2021), such as Anthropic's Constitutional AI (Bai et al 2022), but addresses some of their key shortcomings: transparency, insularity, and accuracy.…”
Section: Overviewmentioning
confidence: 97%
“…Conversely, if the context shifts to "for the purpose of surveillance or spying," the same action loses its moral grounding. This phenomenon of flexibly bending moral rules in instantiations of scenarios is widely recognized in assorted cognitive science studies (Kwon et al, 2022;Levine et al, 2020;Awad et al, 2022;Levine et al, 2018).…”
Section: Introductionmentioning
confidence: 94%