The 2011 International Joint Conference on Neural Networks 2011
DOI: 10.1109/ijcnn.2011.6033519
|View full text |Cite
|
Sign up to set email alerts
|

Modeling a system for monitoring an object using artificial neural networks and reinforcement learning

Abstract: This paper presents a modeling of a system designed to monitor a moving object from images captured by a camera. The research was focused on defining the steps necessary to the functioning of systems, they are: capture and image processing, pattern recognition with artificial neural networks and seek the best path for moving the camera, using reinforcement learning. The results show the viability of the proposed system, being a relevant alternative to monitoring and security environments.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2013
2013
2017
2017

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 10 publications
(16 reference statements)
0
2
0
Order By: Relevance
“…The goal of the RL method is to guide the agent towards taking actions that would result in maximizing (or minimizing) the sum of the reinforcement signals (numerical reward or punishment) received over the course of time, known as the expected return, which does not always signify maximizing the immediate reinforcement to be received [ 34 ].…”
Section: Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…The goal of the RL method is to guide the agent towards taking actions that would result in maximizing (or minimizing) the sum of the reinforcement signals (numerical reward or punishment) received over the course of time, known as the expected return, which does not always signify maximizing the immediate reinforcement to be received [ 34 ].…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…The behavior that the agent should adopt in order to achieve maximization (or minimization) of the return is known as the policy and can be expressed by π. According to [ 34 ], a policy π (s,a) is a mapping of states ( s ) in actions ( a ) taken in that state, and represents the probability of selecting each one of the possible actions, in such a way that the best actions correspond to the greatest probabilities of selection. When this mapping maximizes the sum of the rewards, the optimum policy has been achieved.…”
Section: Reinforcement Learningmentioning
confidence: 99%