2018 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton) 2018
DOI: 10.1109/allerton.2018.8635871
|View full text |Cite
|
Sign up to set email alerts
|

Deception in Optimal Control

Abstract: In this paper, we consider an adversarial scenario where one agent seeks to achieve an objective and its adversary seeks to learn the agent's intentions and prevent the agent from achieving its objective. The agent has an incentive to try to deceive the adversary about its intentions, while at the same time working to achieve its objective. The primary contribution of this paper is to introduce a mathematically rigorous framework for the notion of deception within the context of optimal control. The central no… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 18 publications
(9 citation statements)
references
References 34 publications
(49 reference statements)
0
9
0
Order By: Relevance
“…Given a known outcome matrix and opponent model, the algorithm selects a communication that maximises the deceiver's outcome for a single action. This algorithm is therefore most similar to deception-maximising approaches [18,19,21,22,26], and is not suitable for modelling strategic deception in OR simulations.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Given a known outcome matrix and opponent model, the algorithm selects a communication that maximises the deceiver's outcome for a single action. This algorithm is therefore most similar to deception-maximising approaches [18,19,21,22,26], and is not suitable for modelling strategic deception in OR simulations.…”
Section: Related Workmentioning
confidence: 99%
“…Recent work on computational deception has focused on 'singular acts of deception' [28], where an agent attempts to maximise the deceptiveness of an action or plan according to a predefined metric [18,19,21,22,26]. These approaches address decision problems for which the deception itself is the goal: success is measured according to the reduced accuracy of the observer's goal recognition model.…”
Section: Introductionmentioning
confidence: 99%
“…4, there are three banks locating in three 1 × 1 blocks in the state space, where the robber's goal is to reach bank 2 while deceiving the cop (as an adversary) to believe that the robber is trying to reach the other two banks (see [14] for details of the game). We simulate 192 trajectories with length 100, with the policies obtained from [14], starting from the 64 different initial states with 3 different initial beliefs of the cop. We set the prior probability distribution as a stationary uniform distribution in the state space.…”
Section: Case Iii: Explaining Policies Of Mdpsmentioning
confidence: 99%
“…Out of the box of cybersecurity, relative researches across deception and stochastic decision-making process are just emerging in optimal control and reinforcement learning areas. Ornik and Topcu in [11] have introduced a mathematical framework for deception in optimal control based on Markov decision process (MDP). However, the framework has a prerequisite of its deception objective being reward-based.…”
Section: Introductionmentioning
confidence: 99%
“…While these researches all demonstrate novel and inspirational work from their respective angle, they are meantime subject to several common limitations. Attacks modeled in researches [11], [23] and [14] are subject to certain "objectives" of deception/attack; research in [21] considers human cognitive bias as a reason for agent's potential suboptimal choices; models in [21], [16], [23] and [14] are computed with underlying MDP regardless of agent's partial observability as a potential setting; and some of the tactics introduced above are more attacking rather than deceptive.…”
Section: Introductionmentioning
confidence: 99%