Abstract
This article studies the problem of data-driven combined longitudinal and lateral control of autonomous vehicles (AVs) such that the AV can stay within a safe but minimum distance from its leading vehicle and, at the same time, in the lane. Most of the existing methods for combined longitudinal and lateral control are either model-based or developed by purely data-driven methods such as reinforcement learning. Traditional model-based control approaches are insufficient to address the adaptive optimal control design issue for AVs in dynamically changing environments and are subject to model uncertainty. Moreover, the conventional reinforcement learning approaches require a large volume of data, and cannot guarantee the stability of the vehicle. These limitations are addressed by integrating the advanced control theory with reinforcement learning techniques. To be more specific, by utilizing adaptive dynamic programming (ADP) techniques and using the motion data collected from the vehicles, a policy iteration algorithm is proposed such that the control policy is iteratively optimized in the absence of the precise knowledge of the AV’s dynamical model. Furthermore, the stability of the AV is guaranteed with the control policy generated at each iteration of the algorithm. The efficiency of the proposed approach is validated by the integrated simulation of SUMO and CommonRoad.
Original language | English (US) |
---|---|
Pages (from-to) | 991-1005 |
Number of pages | 15 |
Journal | IEEE Transactions on Control Systems Technology |
Volume | 33 |
Issue number | 3 |
DOIs | |
State | Published - 2025 |
Keywords
- Adaptive dynamic programming (ADP)
- combined longitudinal and lateral control
- connected vehicles
ASJC Scopus subject areas
- Control and Systems Engineering
- Electrical and Electronic Engineering