Our system is currently under heavy load due to increased usage. We're actively working on upgrades to improve performance. Thank you for your patience.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop) 2023
DOI: 10.18653/v1/2023.acl-srw.40
|View full text |Cite
|
Sign up to set email alerts
|

Moral Mimicry: Large Language Models Produce Moral Rationalizations Tailored to Political Identity

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 0 publications
0
1
0
Order By: Relevance
“…Researchers have introduced harm taxonomies specifically for LLMs, which identify known risks (i.e., informed by observed instances of harm) [18,100,190] and emerging risks of LLMs (anticipated risks based on foreseeable capabilities of LLMs) [108,166]. Since LLMs can be used for a wide range of tasks associated with many different categories of harms, researchers have presented frameworks and evaluation methods to assess a particular type of LLM harm, including misinformation [74,135], representation and toxicity [42,64], human autonomy [65,168], malicious use [38,154], and data privacy [87,97]. The popular methods to identify these harms include benchmarking [27,28], user research [101,106], and adversarial testing [41,137].…”
Section: Identifying and Mitigating Llm Harmsmentioning
confidence: 99%
“…Researchers have introduced harm taxonomies specifically for LLMs, which identify known risks (i.e., informed by observed instances of harm) [18,100,190] and emerging risks of LLMs (anticipated risks based on foreseeable capabilities of LLMs) [108,166]. Since LLMs can be used for a wide range of tasks associated with many different categories of harms, researchers have presented frameworks and evaluation methods to assess a particular type of LLM harm, including misinformation [74,135], representation and toxicity [42,64], human autonomy [65,168], malicious use [38,154], and data privacy [87,97]. The popular methods to identify these harms include benchmarking [27,28], user research [101,106], and adversarial testing [41,137].…”
Section: Identifying and Mitigating Llm Harmsmentioning
confidence: 99%
“…Foundation models in particular can increase the scale and speed at which disinformation campaigns can be disseminated across the information ecosystem [161,198]. As generative AI applications powered by foundation models flood the public sphere with fake information, there is a risk of eroding public trust in the information that circulates online, further fueling social polarization and the creation of echo chambers [113,195].…”
Section: Social Risks and Harmsmentioning
confidence: 99%
“…Given the increasing societal role played by Large Language Models (LLMs), researchers have begun to investigate the underlying psychology of these generative models. For example, several works have investigated whether LLMs can truly understand language and perform reasoning (Chowdhery et al, 2022), understand distinctions between different moralities and personalities (Miotto et al, 2022;Simmons, 2022), and learn ethical dilemmas (Jiang et al, 2021). Hagendorff et al (2022), for instance, demonstrated that LLMs are intuitive decision makers, just like humans, arguing that investigating LLMs with methods from psychology has the potential to uncover their emergent traits and behavior.…”
Section: Introductionmentioning
confidence: 99%

Which Humans?

Atari,
Xue,
Park
et al. 2023
Preprint