Abstract
This paper presents a supervised reinforcement learning (SRL)-based framework for longitudinal vehicle dynamics control of cooperative adaptive cruise control (CACC) system. A supervisor network trained by real driving data is incorporated into the actor-critic reinforcement learning approach. In the SRL training process, the actor and critic network are updated under the guidance of the supervisor and the gain scheduler. As a result, the training success rate is improved, and the driver characteristics can be learned by the actor to achieve a human-like CACC controller. The SRL-based control policy is compared with a linear controller in typical driving situations through simulation, and the control policies trained by drivers with different driving styles are compared using a real driving cycle. Furthermore, the proposed control strategy is demonstrated by a real vehicle-following experiment with different time headways. The simulation and experimental results not only validate the effectiveness and adaptability of the SRL-based CACC system, but also show that it can provide natural following performance like human driving.
Original language | English |
---|---|
Article number | 1014 |
Journal | Applied Sciences (Switzerland) |
Volume | 8 |
Issue number | 7 |
DOIs | |
Publication status | Published - 21 Jun 2018 |
Keywords
- Cooperative adaptive cruise control
- Driving behavior
- Longitudinal dynamics control
- Supervised reinforcement learning
- Vehicle following control