Reinforcement Learning for Multi-Agent Systems with an Application to Distributed Predictive Cruise Control

Mohammed Mynuddin, Weinan Gao, Zhong Ping Jiang

Research output: Chapter in Book/Report/Conference proceedingConference contribution


In this paper, we propose a reinforcement learning (RL) approach to the coordinated control of multi-agent systems with a special emphasis on intelligent transportation systems. As a result, for a network of autonomous vehicles, a novel distributed predictive cruise control (PCC) algorithm based on RL is proposed to reduce idle time and maintain an adjustable speed depending on the traffic signals. Under the proposed distributed PCC law, given the signal phase and timing (SPaT) message from upcoming traffic intersections, autonomous vehicles can cross intersections without stopping. The effectiveness of the proposed approach has been validated through Paramics microscopic traffic simulations by choosing a scenario in Statesboro, Georgia. For different traffic demands, the travel time and fuel consumption rate of vehicles are compared between non-PCC and PCC algorithms. Microscopic traffic simulation results show that the proposed PCC algorithm is able to reduce both fuel consumption rate and travel time.

Original languageEnglish (US)
Title of host publication2020 American Control Conference, ACC 2020
PublisherInstitute of Electrical and Electronics Engineers Inc.
Number of pages6
ISBN (Electronic)9781538682661
StatePublished - Jul 2020
Event2020 American Control Conference, ACC 2020 - Denver, United States
Duration: Jul 1 2020Jul 3 2020

Publication series

NameProceedings of the American Control Conference
ISSN (Print)0743-1619


Conference2020 American Control Conference, ACC 2020
Country/TerritoryUnited States

ASJC Scopus subject areas

  • Electrical and Electronic Engineering


Dive into the research topics of 'Reinforcement Learning for Multi-Agent Systems with an Application to Distributed Predictive Cruise Control'. Together they form a unique fingerprint.

Cite this