2021
DOI: 10.1007/s40313-021-00786-x
|View full text |Cite
|
Sign up to set email alerts
|

An Improved Q-learning Approach with Kalman Filter for Self-balancing Robot Using OpenAI

Abstract: A two-wheeled self-balancing robot (SBR) is a typical example in control systems that works on the principle of an inverted pendulum. In this paper, we experiment to see how the learning and stability performance varies based on Kalman filter introduction for IMU noise filtering and controlling the robot using reinforcement learning. All the implementation is performed in ROS and Gazebo, and Q-learning is implemented using OpenAI (toolkit for development of Reinforcement learning) for ROS, i.e., Openai_ros pac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 8 publications
0
6
0
Order By: Relevance
“…Recently, the authors in Reference 18 published research involving the use of KF to improve reinforcement learning applied to the control of a two‐wheeled robot, similar to the inverted pendulum problem. Different from our proposal, which uses EKF as a predictor of future robot states, the authors apply KF to filter out noisy information in sensory data before being used as a state in Q‐Learning.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, the authors in Reference 18 published research involving the use of KF to improve reinforcement learning applied to the control of a two‐wheeled robot, similar to the inverted pendulum problem. Different from our proposal, which uses EKF as a predictor of future robot states, the authors apply KF to filter out noisy information in sensory data before being used as a state in Q‐Learning.…”
Section: Related Workmentioning
confidence: 99%
“…This technique has been shown to improve the accuracy of states measured by a sensor, in this case the IMU (inertial measurement unit), thus it increases the stabilization of the obtained rewards and the transient response of the control system. Therefore, from the results presented in Reference 18, we are led to investigate if in addition to providing greater stabilization, KF can accelerate a newer or current reinforcement learning technique such as DQN.…”
Section: Related Workmentioning
confidence: 99%
“…The true values of the weight parameters of the deep Q-network may have some deviation or fluctuation from the values we train, which may come from factors such as noise in the training data, randomness of the optimization algorithm, and complexity of the network structure. The uncertainty of network parameters can affect the performance and stability of the network, so it is necessary to analyze and filter out the uncertainty of network parameters [16].…”
Section: Dqn-ekf Algorithmmentioning
confidence: 99%
“…Recently, the authors in 16 published research involving the use of KF to improve reinforcement learning applied to the control of a two-wheeled robot, similar to the inverted pendulum problem. Different from our proposal, which uses EKF as a predictor of future robot states, the authors apply KF to filter out noisy information in sensory data before being used as a state in Q-Learning.…”
Section: Related Workmentioning
confidence: 99%
“…This technique has been shown to improve the accuracy of states measured by a sensor, in this case the IMU (Inertial Measurement Unit), thus it increases the stabilization of the obtained rewards and the transient response of the control system. Therefore, from the results presented in 16 , we are led to investigate if in addition to providing greater stabilization, KF can accelerate a newer or current reinforcement learning technique such as DQN.…”
Section: Related Workmentioning
confidence: 99%