Decentralized Multi-Agent Reinforcement Learning with Visible Light Communication for Robust Urban Traffic Signal Control
Manuel Augusto Vieira,
Gonçalo Galvão,
Manuela Vieira (),
Mário Véstias,
Paula Louro and
Pedro Vieira
Additional contact information
Manuel Augusto Vieira: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Gonçalo Galvão: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Manuela Vieira: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Mário Véstias: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Paula Louro: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Pedro Vieira: DEETC-ISEL/IPL, R. Conselheiro Emídio Navarro, 1949-014 Lisboa, Portugal
Sustainability, 2025, vol. 17, issue 22, 1-32
Abstract:
The rapid growth of urban vehicle and pedestrian flows has intensified congestion, delays, and safety concerns, underscoring the need for sustainable and intelligent traffic management in modern cities. Traditional centralized traffic signal control systems often face challenges of scalability, heterogeneity of traffic patterns, and limited real-time adaptability. To address these limitations, this study proposes a decentralized Multi-Agent Reinforcement Learning (MARL) framework for adaptive traffic signal control, where Deep Reinforcement Learning (DRL) agents are deployed at each intersection and trained on local conditions to enable real-time decision-making for both vehicles and pedestrians. A key innovation lies in the integration of Visible Light Communication (VLC), which leverages existing LED-based infrastructure in traffic lights, streetlights, and vehicles to provide high-capacity, low-latency, and energy-efficient data exchange, thereby enhancing each agent’s situational awareness while promoting infrastructure sustainability. The framework introduces a queue–request–response mechanism that dynamically adjusts signal phases, resolves conflicts between flows, and prioritizes urgent or emergency movements, ensuring equitable and safer mobility for all users. Validation through microscopic simulations in SUMO and preliminary real-world experiments demonstrates reductions in average waiting time, travel time, and queue lengths, along with improvements in pedestrian safety and energy efficiency. These results highlight the potential of MARL–VLC integration as a sustainable, resilient, and human-centered solution for next-generation urban traffic management.
Keywords: sustainable urban mobility; intelligent traffic management; multi-agent reinforcement learning (MARL); deep reinforcement learning (DRL); visible light communication (VLC); energy efficiency; pedestrian safety; smart cities (search for similar items in EconPapers)
JEL-codes: O13 Q Q0 Q2 Q3 Q5 Q56 (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2071-1050/17/22/10056/pdf (application/pdf)
https://www.mdpi.com/2071-1050/17/22/10056/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jsusta:v:17:y:2025:i:22:p:10056-:d:1791849
Access Statistics for this article
Sustainability is currently edited by Ms. Alexandra Wu
More articles in Sustainability from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().