| Citation: | LEI Y L,DING W R,LUO Y Z,et al. Trajectory planning and resource allocation optimization in UAV data collection missions[J]. Journal of Beijing University of Aeronautics and Astronautics,2025,51(10):3460-3470 (in Chinese) doi: 10.13700/j.bh.1001-5965.2023.0531 |
A joint optimization method for unmanned aerial vehicle (UAV) trajectory planning and resource allocation based on deep reinforcement learning was proposed to address the challenges of limited battery capacity, limited cache space, and dynamic changes in ground target priorities during data collection tasks in emergency scenarios. First, a mathematical model was developed by considering the communication, computation, flight, and data caching processes in UAV missions. Then, a Markov process model was established for UAV trajectory planning and resource allocation, with corresponding state and action descriptions. A weighted reward function was designed to balance UAV energy consumption and data collection volume. Finally, simulations were conducted to compare the proposed method with greedy algorithms and genetic algorithms. The results show that the proposed method can significantly improve the amount of data collected from ground users within a shorter task time, at a similar or lower energy cost for UAVs.
| [1] |
胡欣颖. 面向信息采集场景的无人机轨迹规划与通信资源分配研究[D]. 北京: 北京邮电大学, 2021: 1-3.
HU X Y. Research on UAV trajectory planning and communication resource allocation for information acquisition scenes[D]. Beijing: Beijing University of Posts and Telecommunications, 2021: 1-3(in Chinese).
|
| [2] |
LI Z R, ZHANG Y W, WU H, et al. Design and application of a UAV autonomous inspection system for high-voltage power transmission lines[J]. Remote Sensing, 2023, 15(3): 865. doi: 10.3390/rs15030865
|
| [3] |
DONG M X, OTA K, LIN M, et al. UAV-assisted data gathering in wireless sensor networks[J]. The Journal of Supercomputing, 2014, 70(3): 1142-1155. doi: 10.1007/s11227-014-1161-6
|
| [4] |
GONG J, CHANG T H, SHEN C, et al. Flight time minimization of UAV for data collection over wireless sensor networks[J]. IEEE Journal on Selected Areas in Communications, 2018, 36(9): 1942-1954. doi: 10.1109/JSAC.2018.2864420
|
| [5] |
CHEN X, CHEN X M. The UAV dynamic path planning algorithm research based on Voronoi diagram[C]//Proceedings of the 26th Chinese Control and Decision Conference. Piscataway: IEEE Press, 2014: 1069-1071.
|
| [6] |
HSU D, LATOMBE J C, KURNIAWATI H. On the probabilistic foundations of probabilistic roadmap planning[J]. International Journal of Robotics Research, 2006, 25(7): 627-643.
|
| [7] |
LIU X. Four alternative patterns of the Hilbert curve[J]. Applied Mathematics and Computation, 2004, 147(3): 741-752. doi: 10.1016/S0096-3003(02)00808-1
|
| [8] |
MU X D, LIU Y W, GUO L, et al. Energy-constrained UAV data collection systems: NOMA and OMA[J]. IEEE Transactions on Vehicular Technology, 2021, 70(7): 6898-6912. doi: 10.1109/TVT.2021.3086556
|
| [9] |
SAMIR M, SHARAFEDDINE S, ASSI C M, et al. UAV trajectory planning for data collection from time-constrained IoT devices[J]. IEEE Transactions on Wireless Communications, 2020, 19(1): 34-46. doi: 10.1109/TWC.2019.2940447
|
| [10] |
ZHAN C, LAI H. Energy minimization in internet-of-things system based on rotary-wing UAV[J]. IEEE Wireless Communications Letters, 2019, 8(5): 1341-1344. doi: 10.1109/LWC.2019.2916549
|
| [11] |
NGUYEN K K, DUONG T Q, DO-DUY T, et al. 3D UAV trajectory and data collection optimisation via deep reinforcement learning[J]. IEEE Transactions on Communications, 2022, 70(4): 2358-2371. doi: 10.1109/TCOMM.2022.3148364
|
| [12] |
FU S, TANG Y J, WU Y, et al. Energy-efficient UAV-enabled data collection via wireless charging: a reinforcement learning approach[J]. IEEE Internet of Things Journal, 2021, 8(12): 10209-10219. doi: 10.1109/JIOT.2021.3051370
|
| [13] |
BAYERLEIN H, THEILE M, CACCAMO M, et al. UAV path planning for wireless data harvesting: a deep reinforcement learning approach[C]//Proceedings of the IEEE Global Communications Conference. Piscataway: IEEE Press, 2020: 1-6.
|
| [14] |
HAARNOJA T, ZHOU A, ABBEEL P, et al. Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor[C]//Proceedings of the International Conference on Machine Learning. Piscataway: IEEE Press, 2018: 1861-1870.
|
| [15] |
ISLAM S K, HAIDER M R. Sensors and low power signal processing[M]. Berlin: Springer Science & Business Media, 2009.
|
| [16] |
LIU Q, SHI L, SUN L L, et al. Path planning for UAV-mounted mobile edge computing with deep reinforcement learning[J]. IEEE Transactions on Vehicular Technology, 2020, 69(5): 5723-5728. doi: 10.1109/TVT.2020.2982508
|
| [17] |
WU Q Q, ZENG Y, ZHANG R. Joint trajectory and communication design for multi-UAV enabled wireless networks[J]. IEEE Transactions on Wireless Communications, 2018, 17(3): 2109-2121. doi: 10.1109/TWC.2017.2789293
|
| [18] |
LAMBORA A, GUPTA K, CHOPRA K. Genetic algorithm-a literature review[C]//Proceedings of the International Conference on Machine Learning, Big Data, Cloud and Parallel Computing. Piscataway: IEEE Press, 2019: 380-384.
|