UAV Control Method Combining Reptile Meta-Reinforcement Learning and Generative Adversarial Imitation Learning
Shui Jiang,
Yanning Ge,
Xu Yang (),
Wencheng Yang and
Hui Cui
Additional contact information
Shui Jiang: College of Computer and Cyber Security, Fujian Normal University, Fuzhou 350007, China
Yanning Ge: College of Computer and Cyber Security, Fujian Normal University, Fuzhou 350007, China
Xu Yang: College of Computer and Control Engineering, Minjiang University, Fuzhou 350108, China
Wencheng Yang: School of Mathematics, Physics and Computing, University of Southern Queensland, Darling Heights, QLD 4350, Australia
Hui Cui: Department of Software Systems & Cybersecurity, Monash University, Melbourne, VIC 3800, Australia
Future Internet, 2024, vol. 16, issue 3, 1-18
Abstract:
Reinforcement learning (RL) is pivotal in empowering Unmanned Aerial Vehicles (UAVs) to navigate and make decisions efficiently and intelligently within complex and dynamic surroundings. Despite its significance, RL is hampered by inherent limitations such as low sample efficiency, restricted generalization capabilities, and a heavy reliance on the intricacies of reward function design. These challenges often render single-method RL approaches inadequate, particularly in the context of UAV operations where high costs and safety risks in real-world applications cannot be overlooked. To address these issues, this paper introduces a novel RL framework that synergistically integrates meta-learning and imitation learning. By leveraging the Reptile algorithm from meta-learning and Generative Adversarial Imitation Learning (GAIL), coupled with state normalization techniques for processing state data, this framework significantly enhances the model’s adaptability. It achieves this by identifying and leveraging commonalities across various tasks, allowing for swift adaptation to new challenges without the need for complex reward function designs. To ascertain the efficacy of this integrated approach, we conducted simulation experiments within both two-dimensional environments. The empirical results clearly indicate that our GAIL-enhanced Reptile method surpasses conventional single-method RL algorithms in terms of training efficiency. This evidence underscores the potential of combining meta-learning and imitation learning to surmount the traditional barriers faced by reinforcement learning in UAV trajectory planning and decision-making processes.
Keywords: unmanned aerial vehicles (UAVs); meta-reinforcement learning; generative adversarial imitation learning (search for similar items in EconPapers)
JEL-codes: O3 (search for similar items in EconPapers)
Date: 2024
References: View references in EconPapers View complete reference list from CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/1999-5903/16/3/105/pdf (application/pdf)
https://www.mdpi.com/1999-5903/16/3/105/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jftint:v:16:y:2024:i:3:p:105-:d:1360431
Access Statistics for this article
Future Internet is currently edited by Ms. Grace You
More articles in Future Internet from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().