2015 10th Asian Control Conference (ASCC) 2015
DOI: 10.1109/ascc.2015.7244682
|View full text |Cite
|
Sign up to set email alerts
|

Cooperative capture by multi-agent using reinforcement learning application for security patrol systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
5
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 1 publication
0
5
0
Order By: Relevance
“…Sugiyama et al [17] designed different visit frequencies for all purposes in the patrolling area and applied divisional cooperation to achieve the patrol tasks. Yasuyuki et al [18] applied reinforcement learning based on discrete patrolling areas to perform security patrol tasks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Sugiyama et al [17] designed different visit frequencies for all purposes in the patrolling area and applied divisional cooperation to achieve the patrol tasks. Yasuyuki et al [18] applied reinforcement learning based on discrete patrolling areas to perform security patrol tasks.…”
Section: Related Workmentioning
confidence: 99%
“…PSO uses a fitness function to evaluate the quality of particles (line 16). The global optimal position of the particle swarm is updated iteratively, which is iterated until the maximum number of iterations (lines [17][18]. must return to the base station and replenish fuel, and then continue to visit the area to be visited.…”
Section: Patrolling Path Planning Algorithm Based On Improved Psomentioning
confidence: 99%
“…The use of robots as a bodyguards is related to several different area of research and has received significant attention lately. Several different studies such as (Richard Klima, 2016;Yasuyuki et al, 2015) considered using robots and multi-agent reinforcement learning for security related tasks such as patrolling and team coordination by placing checkpoints to provide protection against imminent adversaries. A multi-robot patrolling framework was proposed by (Khan et al, 2012) that analyzes the behavior pattern of the soldiers and the robot and generates a patrolling schedule.…”
Section: Related Workmentioning
confidence: 99%
“…[9] To solve the problem regarding the identification of the best strategy to enclose an intruder for pursuers and adopt the huge environment,SHIMADA Yasuyuki and OHTSUKA Hirofumi discussed how to discretize patrol areas based on reinforcement learning. [10] Francisco Martinez-Gil and Miguel Lozano compared the two different learning algorithms based on Reinforcement learning :Iterative Vector Quantization with Q-Learning (ITVQQL) TS, Tile coding as the generalization method with the Sarsa(TS). The simulation results for pedestrian showed that the two RL framework had common advantage that they could generate emergent collective behavior which can help them arrive their destination collaborative.…”
Section: Related Workmentioning
confidence: 99%
“…GetResponse(X) The coalitions are short-lived and goal-oriented [6].If a coalition lasts longer than the default requirement lif e, it will be considered as a failure. The algorithms shown in Algorithm 3 are explained in the following manner:The organizer initializes the neuron network(01).then the pursuers start to catch evaders under the coordination of organizers until all evaders are captured(02-33).The organizer broadcast position and value of each evader found in environment and wait the response from pursuers(03-06).The pursuers response with their own feature vector − → x p (07-09).The organizer train the SOM layer with the set of feature vectors X and send target and − −− → GAF to each pursuer after creating group p (10)(11)(12)(13)(14).The pursuer p response their CEF (p) and start get close to its target (15)(16)(17)(18)(19)(20)(21)(22)(23)(24)(25)(26)(27)(28)(29)(30).The organizer train its feature extraction part with − −− → CEF (31-32).The sequences diagram describing orginaser and pursuer communications is shown in Figure13.…”
mentioning
confidence: 99%