2009 IEEE International Conference on Robotics and Automation 2009
DOI: 10.1109/robot.2009.5152834
|View full text |Cite
|
Sign up to set email alerts
|

Model-based and model-free reinforcement learning for visual servoing

Abstract: Abstract-To address the difficulty of designing a controller for complex visual-servoing tasks, two learning-based uncalibrated approaches are introduced. The first method starts by building an estimated model for the visual-motor forward kinematic of the vision-robot system by a locally linear regression method. Afterwards, it uses a reinforcement learning method named Regularized Fitted Q-Iteration to find a controller (i.e. policy) for the system (model-based RL). The second method directly uses samples com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2010
2010
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 27 publications
(13 citation statements)
references
References 26 publications
0
13
0
Order By: Relevance
“…Regression-based neural networks (NNs) have been considerably used to approximate the non-linear image-Jacobian. [25][26][27][28] They suffer from local minima which are complex to avoid. 29 It is worth mentioning that using NN in approximating the hybrid interaction matrices is not widely investigated in the literature.…”
Section: Related Workmentioning
confidence: 99%
“…Regression-based neural networks (NNs) have been considerably used to approximate the non-linear image-Jacobian. [25][26][27][28] They suffer from local minima which are complex to avoid. 29 It is worth mentioning that using NN in approximating the hybrid interaction matrices is not widely investigated in the literature.…”
Section: Related Workmentioning
confidence: 99%
“…Traditionally, model predictive control (MPC) relies on known or learned dynamics robot models, and it is used to generate advanced motor controls. Once the model of the effect of the robot's actions on its environment (e.g., workspace with objects) is known or learned, it can be used to plan a sequence of optimal actions to fulfill a task: this can be done with MPC, often with a receding horizon approach, but also with reinforcement learning [25], [26], [27]. For planning sequence of manipulation actions, this was often done with reduced visual models computing features of the objects in the scene, to lower the dimensionality of the problem.…”
Section: Related Workmentioning
confidence: 99%
“…The second ingredient for UVS seems to be available and easy to implement (for example the Broyden update approach), thanks to extensive research and evaluation in control [10], [13], [14]. However the first one, which is the crucial one, despite a lot of research [15], [16], [17], [18], is still hard to synthesize for reliable tracking of everyday objects as shown by Fig.…”
Section: Visual Servoing In Practicementioning
confidence: 99%