Deep reinforcement learning for dynamic incident-responsive traffic information dissemination
This study is concerned with the optimal dynamical information dissemination (DID) problem in a transportation network interrupted by traffic incidents. Optimizing system performance with DID after road incidents is challenging because of the uncertainty in traffic flow variation and travelers’ hete...
Gespeichert in:
Veröffentlicht in: | Transportation research. Part E, Logistics and transportation review Logistics and transportation review, 2022-10, Vol.166, p.102871, Article 102871 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | This study is concerned with the optimal dynamical information dissemination (DID) problem in a transportation network interrupted by traffic incidents. Optimizing system performance with DID after road incidents is challenging because of the uncertainty in traffic flow variation and travelers’ heterogeneous responses to information. To address the problem, we consider a traffic manager who aims to improve the system performance by dynamically generating and disseminating information to road users in a time period after an incident happens. We develop a decision tool for obtaining DID strategy based on double deep Q-learning (DDQL) for the traffic manager, aiming at finding an optimal DID strategy. The decision tool is integrated with traffic sensors which collect traffic data in real time. With advanced traveler information systems, the DID system dynamically sends out various types of information to users according to the current and anticipated traffic states so as to minimize congestion and enhance road network capacity. In particular, the proposed DDQL method utilizes a double deep Q-network (DQN) structure to learn the state–action values. To test and evaluate the performance of the decision tool, we develop a microscopic simulation model of a real road network in the Serangoon area of Singapore in PTV VISSIM and calibrate the model with real historical traffic data. We train and compare the DDQL controller model with different reward signals, including the weighted sum of the average speed and queue delay, total traffic flow, and average travel time. Numerical experiments demonstrate the good performance of the proposed DDQL-based DID approach in improving the congestion and other performance metrics of the expressway. The robustness and generalizability of the DDQL agent are also verified by evaluating the algorithm performance in environments with different traffic demand patterns and driving behavior profiles.
•Develop a DDQL framework to generate optimal dynamic information dissemination policies.•Swiftly disseminate information to road users at multiple locations•Consider the uncertainty in travelers’ aggregate diversion decisions.•Evaluate policy performance with different reward signals.•Demonstrate the performance of our solution framework by comparing it with various benchmarks. |
---|---|
ISSN: | 1366-5545 1878-5794 |
DOI: | 10.1016/j.tre.2022.102871 |