Abstract:In the light of ongoing progresses of research on artificial intelligent systems exhibiting a steadily increasing problem-solving ability, the identification of practicable solutions to the value alignment problem in AGI Safety is becoming a matter of urgency. In this context, one preeminent challenge that has been addressed by multiple researchers is the adequate formulation of utility functions or equivalents reliably capturing human ethical conceptions. However, the specification of suitable utility functio… Show more
“…Thereby, for safety reasons, the utility functions can and should include context-sensitive and perceiver-dependent elements as integrated e.g. in augmented utilitarianism [13]. Fourth, updates of law are solely reflected in the ethical goal functions which leads to a more flexible and controllable task.…”
Section: Disentanglement Of Responsibilitiesmentioning
confidence: 99%
“…Fifth, a blockchain approach to ensure the security and transparency of the goal functions themselves and all updates on these functions might be recommendable. Crucially, in order to avoid formulations of an ethical goal function with safety-critical side effects for human entities (including implications related to impossibility theorems for consequentialist frameworks [152]), it is recommendable to assign a type of perceiverdependent and context-sensitive utility to simulations of situations instead of only to the future outcome of actions [14,13]. In the long-term, we believe that scientific research with the goal to integrate the first-person perspective of society on perceived well-being within an ethical goal function at the core of the presented socio-technological feedbackloop might represent one substantial element needed to promote human flourishing in the most efficient possible way aided by the problem solving ability of AI.…”
Section: Conclusion and Future Prospectsmentioning
confidence: 99%
“…From an abstract point of view, one could distinguish different means by which an adversary might achieve successful attacks: e.g. 1) by fooling the AI at the perception-level (in analogy to classical adversarial examples [198], this variant has been denoted ethical adversarial examples [13]) which could lead to an unethical behavior even if the utility function would have been aligned with human ethical Figure 6.1: Intuitive illustration for the law of requisite variety. Taken from [320].…”
Section: Variety In Embodied Moralitymentioning
confidence: 99%
“…states (of affairs) s . In line with this insight, is the context-sensitive and perceiverdependent type of utility functions considering agent, action and outcome which has been recently proposed within a novel ethical framework denoted augmented utilitarianism [13] (abbreviated with AU in the following) which was introduced in Chapter 4. Reconsidering the dyadic morality template iA d − → vP , it seems that in order to better capture the variety of human morality, utility functions -now transferring it to the perspective of Type I AI systems -would need to be at least formulated at the abstraction level of a perceiverdependent evaluation of a transition s a − → s leading from a state s to a state s via an action a.…”
Section: Variety Through "Dyadicness"mentioning
confidence: 99%
“…For the purpose of facilitating the injection of requisite knowledge reflecting the variety of human morality in ethical goal functions, Section 6.2 provides information on the following variety-relevant aspects: 1) the essential role of affect and emotion in moral judgements from a modern constructionist neuroscience and cognitive science perspective followed by 2) dyadic morality as a recent psychological theory on the nature of cognitive templates for moral judgements. In Section 6.3, we propose first guidelines on how to approximately formulate ethical goal functions using a recently proposed non-normative socio-technological ethical framework grounded in science called augmented utilitarianism [13] that might be useful to better incorporate the requisite variety of human ethical intuitions (especially in comparison to classical utilitarianism). Thereafter, we propose how to possibly validate these functions within a socio-technological feedback-loop [14].…”
As the first author of the underlying paper, I had a vital contribution and it was solely my responsibility to write down the content and to perform an extensive literature research as well as in-depth analysis. This chapter is based on a slightly modified form of the publication: N.-M. Aliman and L. Kester. Augmented Utilitarianism for AGI Safety. In International Conference on Artificial General Intelligence, pages 11-21. Springer, 2019. As the first author of the underlying paper, I had a vital contribution and it was solely my responsibility to write down the content and to perform an extensive literature research as well as in-depth analysis.
“…Thereby, for safety reasons, the utility functions can and should include context-sensitive and perceiver-dependent elements as integrated e.g. in augmented utilitarianism [13]. Fourth, updates of law are solely reflected in the ethical goal functions which leads to a more flexible and controllable task.…”
Section: Disentanglement Of Responsibilitiesmentioning
confidence: 99%
“…Fifth, a blockchain approach to ensure the security and transparency of the goal functions themselves and all updates on these functions might be recommendable. Crucially, in order to avoid formulations of an ethical goal function with safety-critical side effects for human entities (including implications related to impossibility theorems for consequentialist frameworks [152]), it is recommendable to assign a type of perceiverdependent and context-sensitive utility to simulations of situations instead of only to the future outcome of actions [14,13]. In the long-term, we believe that scientific research with the goal to integrate the first-person perspective of society on perceived well-being within an ethical goal function at the core of the presented socio-technological feedbackloop might represent one substantial element needed to promote human flourishing in the most efficient possible way aided by the problem solving ability of AI.…”
Section: Conclusion and Future Prospectsmentioning
confidence: 99%
“…From an abstract point of view, one could distinguish different means by which an adversary might achieve successful attacks: e.g. 1) by fooling the AI at the perception-level (in analogy to classical adversarial examples [198], this variant has been denoted ethical adversarial examples [13]) which could lead to an unethical behavior even if the utility function would have been aligned with human ethical Figure 6.1: Intuitive illustration for the law of requisite variety. Taken from [320].…”
Section: Variety In Embodied Moralitymentioning
confidence: 99%
“…states (of affairs) s . In line with this insight, is the context-sensitive and perceiverdependent type of utility functions considering agent, action and outcome which has been recently proposed within a novel ethical framework denoted augmented utilitarianism [13] (abbreviated with AU in the following) which was introduced in Chapter 4. Reconsidering the dyadic morality template iA d − → vP , it seems that in order to better capture the variety of human morality, utility functions -now transferring it to the perspective of Type I AI systems -would need to be at least formulated at the abstraction level of a perceiverdependent evaluation of a transition s a − → s leading from a state s to a state s via an action a.…”
Section: Variety Through "Dyadicness"mentioning
confidence: 99%
“…For the purpose of facilitating the injection of requisite knowledge reflecting the variety of human morality in ethical goal functions, Section 6.2 provides information on the following variety-relevant aspects: 1) the essential role of affect and emotion in moral judgements from a modern constructionist neuroscience and cognitive science perspective followed by 2) dyadic morality as a recent psychological theory on the nature of cognitive templates for moral judgements. In Section 6.3, we propose first guidelines on how to approximately formulate ethical goal functions using a recently proposed non-normative socio-technological ethical framework grounded in science called augmented utilitarianism [13] that might be useful to better incorporate the requisite variety of human ethical intuitions (especially in comparison to classical utilitarianism). Thereafter, we propose how to possibly validate these functions within a socio-technological feedback-loop [14].…”
As the first author of the underlying paper, I had a vital contribution and it was solely my responsibility to write down the content and to perform an extensive literature research as well as in-depth analysis. This chapter is based on a slightly modified form of the publication: N.-M. Aliman and L. Kester. Augmented Utilitarianism for AGI Safety. In International Conference on Artificial General Intelligence, pages 11-21. Springer, 2019. As the first author of the underlying paper, I had a vital contribution and it was solely my responsibility to write down the content and to perform an extensive literature research as well as in-depth analysis.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.