Deep Reinforcement Learning for Real-Time Airport Emergency Evacuation Using Asynchronous Advantage Actor–Critic (A3C) Algorithm
Yujing Zhou,
Yupeng Yang,
Bill Deng Pan,
Yongxin Liu,
Sirish Namilae,
Houbing Herbert Song and
Dahai Liu ()
Additional contact information
Yujing Zhou: Department of Electrical Engineering and Computer Science, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA
Yupeng Yang: College of Engineering, University of North Carolina at Charlotte, Charlotte, NC 28223, USA
Bill Deng Pan: School of Graduate Studies in College of Aviation, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA
Yongxin Liu: Department of Mathematics in College of Arts and Sciences, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA
Sirish Namilae: Department of Aerospace Engineering, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA
Houbing Herbert Song: College of Engineering and Information Technology, University of Maryland, Baltimore County, Baltimore, MD 21250, USA
Dahai Liu: School of Graduate Studies in College of Aviation, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA
Mathematics, 2025, vol. 13, issue 14, 1-22
Abstract:
Emergencies can occur unexpectedly and require immediate action, especially in aviation, where time pressure and uncertainty are high. This study focused on improving emergency evacuation in airport and aircraft scenarios using real-time decision-making support. A system based on the Asynchronous Advantage Actor–Critic (A3C) algorithm, an advanced deep reinforcement learning method, was developed to generate faster and more efficient evacuation routes compared to traditional models. The A3C model was tested in various scenarios, including different environmental conditions and numbers of agents, and its performance was compared with the Deep Q-Network (DQN) algorithm. The results showed that A3C achieved evacuations 43.86% faster on average and converged in fewer episodes (100 vs. 250 for DQN). In dynamic environments with moving threats, A3C also outperformed DQN in maintaining agent safety and adapting routes in real time. As the number of agents increased, A3C maintained high levels of efficiency and robustness. These findings demonstrate A3C’s strong potential to enhance evacuation planning through improved speed, adaptability, and scalability. The study concludes by highlighting the practical benefits of applying such models in real-world emergency response systems, including significantly faster evacuation times, real-time adaptability to evolving threats, and enhanced scalability for managing large crowds in high-density environments including airport terminals. The A3C-based model offers a cost-effective alternative to full-scale evacuation drills by enabling virtual scenario testing, supports proactive safety planning through predictive modeling, and contributes to the development of intelligent decision-support tools that improve coordination and reduce response time during emergencies.
Keywords: emergency evacuation; aviation safety; reinforcement learning; A3C algorithm; DQN algorithm; airport simulation; real-time decision-making (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/13/14/2269/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/14/2269/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:14:p:2269-:d:1701636
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().