2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC) 2019
DOI: 10.1109/ccwc.2019.8666613
|View full text |Cite
|
Sign up to set email alerts
|

Reverse Parking a Car-Like Mobile Robot with Deep Reinforcement Learning and Preview Control

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…Different setups for DQN-based motion planning [ 7 ] have been compared, which demonstrated its acceptable runtime performance on several devices. The deep determinist policy gradient (DDPG) [ 9 ] with preview control, which relies on a reference signal, has been proposed to solve the optimal control problem of vertical parking. DDPG [ 8 ] with manual guide exploration and different control cycle re-training pipelines has been used to achieve reactive end-to-end parking in a real vehicle platform.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Different setups for DQN-based motion planning [ 7 ] have been compared, which demonstrated its acceptable runtime performance on several devices. The deep determinist policy gradient (DDPG) [ 9 ] with preview control, which relies on a reference signal, has been proposed to solve the optimal control problem of vertical parking. DDPG [ 8 ] with manual guide exploration and different control cycle re-training pipelines has been used to achieve reactive end-to-end parking in a real vehicle platform.…”
Section: Related Workmentioning
confidence: 99%
“…For the original policy iteration, all the trajectories are used to evaluate the policy. In the field of model-free RL, the policy gradient method [ 8 , 9 ] takes the partial derivatives of the expected return of the network parameter such that the return is maximized. Inspired by this, if a new distribution p′ ( τ k ) considering the trajectory return is applied to τ , the new total expected return might be higher [ 30 ]: …”
Section: Data-efficient Rl Algorithm Designmentioning
confidence: 99%
See 1 more Smart Citation
“…However, it requires all behaviors to be repeatedly sampled and the action value function to be discrete. Deep Q-Learning (DQN), proposed in [ 11 , 12 , 13 ], introduces high-dimensional perception into reinforcement learning through deep learning, combining convolutional neural networks with reinforcement learning. DQN can address discrete, low-dimensional action spaces.…”
Section: Introductionmentioning
confidence: 99%
“…The large number of actions makes it difficult to achieve an effective search and the discretization process can lead to information loss. Reference [ 13 ] proposed deterministic policy gradient (DPG) algorithms to demonstrate that deterministic gradient algorithms are more effective than stochastic gradient algorithms and prove their existence. Deep deterministic policy gradient (DDPG) algorithms are built on DPG using a depth function to approximate and learn the policy such that it can be applied in a high-dimensional, continuous action space [ 14 ].…”
Section: Introductionmentioning
confidence: 99%