Distributed agent-based deep reinforcement learning for large scale traffic signal control
File version
Author(s)
Wu, Jianqing
Shen, Jun
Du, Bo
Telikani, Akbar
Fahmideh, Mahdi
Liang, Chao
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
Size
File type(s)
Location
License
Abstract
Traffic signal control (TSC) is an established yet challenging engineering solution that alleviates traffic congestion by coordinating vehicles’ movements at road intersections. Theoretically, reinforcement learning (RL) is a promising method for adaptive TSC in complex urban traffic networks. However, current TSC systems still rely heavily on simplified rule-based methods in practice. In this paper, we propose: (1) two game theory-aided RL algorithms leveraging Nash Equilibrium and RL, namely Nash Advantage Actor–Critic (Nash-A2C) and Nash Asynchronous Advantage Actor–Critic (Nash-A3C); (2) a distributed computing Internet of Things (IoT) architecture for traffic simulation, which is more suitable for distributed TSC methods like the Nash-A3C deployment in its fog layer. We apply both methods in our computing architecture and obtain better performance than benchmark TSC methods by 22.1% and 9.7% reduction of congestion time and network delay, respectively.
Journal Title
Knowledge-Based Systems
Conference Title
Book Title
Edition
Volume
241
Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
Item Access Status
Note
Access the data
Related item(s)
Subject
Artificial intelligence
Data management and data science
Machine learning
Science & Technology
Technology
Computer Science, Artificial Intelligence
Computer Science
Traffic signal control
Persistent link to this record
Citation
Wu, Q; Wu, J; Shen, J; Du, B; Telikani, A; Fahmideh, M; Liang, C, Distributed agent-based deep reinforcement learning for large scale traffic signal control, Knowledge-Based Systems, 2022, 241, pp. 108304