EconPapers    
Economics at your fingertips  
 

Uav trajectory optimization for maximizing the ToI-based data utility in wireless sensor networks

Qing Zhao, Zhen Li, Jianqiang Li, Jianxiong Guo, Xingjian Ding () and Deying Li
Additional contact information
Qing Zhao: Beijing University of Technology
Zhen Li: Beijing University of Technology
Jianqiang Li: Beijing University of Technology
Jianxiong Guo: Beijing Normal University
Xingjian Ding: Beijing University of Technology
Deying Li: Renmin University of China

Journal of Combinatorial Optimization, 2025, vol. 49, issue 3, No 10, 25 pages

Abstract: Abstract It’s a promising way to use Unmanned Aerial Vehicles (UAVs) as mobile base stations to collect data from sensor nodes, especially for large-scale wireless sensor networks. There are a lot of works that focus on improving the freshness of the collected data or the data collection efficiency by scheduling UAVs. Given that sensing data in certain applications is time-sensitive, with its value diminishing as time progresses based on Timeliness of Information (ToI), this paper delves into the UAV Trajectory optimization problem for Maximizing the ToI-based data utility (TMT). We give the formal definition of the problem and prove its NP-Hardness. To solve the TMT problem, we propose a deep reinforcement learning-based algorithm that combines the Action Rejection Mechanism and the Deep Q-Network with Priority Experience Replay (ARM-PER-DQN). Where the action rejection mechanism could reduce the action space and PER helps improve the utilization of experiences with high value, thus increasing the training efficiency. To avoid the unbalanced data collection problem, we also investigate a variant problem of TMT (named V-TMT), i.e., each sensor node can be visited by the UAV at most once. We prove that the V-TMT problem is also NP-Hard, and propose a 2-approximation algorithm as the baseline of the ARM-PER-DQN algorithm. We conduct extensive simulations for the two problems to validate the performance of our designs, and the results show that our ARM-PER-DQN algorithm outperforms other baselines, especially in the V-TMT problem, the ARM-PER-DQN algorithm always outperforms the proposed 2-approximation algorithm, which suggests the effectiveness of our algorithm.

Keywords: Data collection; Trajectory optimization; Deep reinforcement learning; Approximation algorithm (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:

Downloads: (external link)
http://link.springer.com/10.1007/s10878-025-01286-3 Abstract (text/html)
Access to the full text of the articles in this series is restricted.

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:spr:jcomop:v:49:y:2025:i:3:d:10.1007_s10878-025-01286-3

Ordering information: This journal article can be ordered from
https://www.springer.com/journal/10878

DOI: 10.1007/s10878-025-01286-3

Access Statistics for this article

Journal of Combinatorial Optimization is currently edited by Thai, My T.

More articles in Journal of Combinatorial Optimization from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().

 
Page updated 2025-04-13
Handle: RePEc:spr:jcomop:v:49:y:2025:i:3:d:10.1007_s10878-025-01286-3