2018
DOI: 10.3390/app8071014
|View full text |Cite
|
Sign up to set email alerts
|

Design and Experimental Validation of a Cooperative Adaptive Cruise Control System Based on Supervised Reinforcement Learning

Abstract: This work presents a supervised reinforcement learning (SRL)-based framework for longitudinal vehicle dynamics' control of the cooperative adaptive cruise control (CACC) system. By incorporating a supervised network trained by real driving data into the actor-critic framework, the training success rate is improved, and the driver characteristics can be learned by the actor to achieve a human-like CACC controller.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
14
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 32 publications
(18 citation statements)
references
References 40 publications
0
14
0
Order By: Relevance
“…Incorporating a supervised network (trained by real driving data) into the actor-critical framework, improves the success rate of training. Furthermore, the driver's characteristics can be learned by the actor to achieve a human-like CACC controller (Wei et al, 2018). There are several validation challenges for adaptive control, including proving convergence over long durations, guaranteeing controller stability, using new tools to compute statistical error bounds, identifying problems in faulttolerant software, and testing in the presence of adaptation (Jacklin et al, 2004).…”
Section: Run-time Monitoring Of Operation Of Control Sub-systemsmentioning
confidence: 99%
“…Incorporating a supervised network (trained by real driving data) into the actor-critical framework, improves the success rate of training. Furthermore, the driver's characteristics can be learned by the actor to achieve a human-like CACC controller (Wei et al, 2018). There are several validation challenges for adaptive control, including proving convergence over long durations, guaranteeing controller stability, using new tools to compute statistical error bounds, identifying problems in faulttolerant software, and testing in the presence of adaptation (Jacklin et al, 2004).…”
Section: Run-time Monitoring Of Operation Of Control Sub-systemsmentioning
confidence: 99%
“…In [18], a longitudinal dynamic model is considered for predictive speed control using DDPG. In [19], a carfollowing controller is developed with acceleration command dynamics considered using DDPG by learning from naturalistic human-driving data. However, both studies did not investigate the impact of acceleration delay, which could degrade the control performance.…”
Section: Introductionmentioning
confidence: 99%
“…Thus, to solve traffic congestion problems, this paper proposes an intelligent traffic light system to initiate a better traffic light cycle configuration and to simulate the schemes by which traffic officers control traffic lights by analyzing the historical data. To achieve this system, a smart system is needed to adjust the cycle configuration of traffic lights by applying reinforcement learning [1][2][3] in machine learning [4].…”
Section: Introductionmentioning
confidence: 99%