Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
Formations of fixed-wing unmanned aerial vehicles (UAVs), which are commonly used in military, rescue, and other missions, often do not have the ability to hover and have a large turning radius. Thus, when operating in an unknown environment, it is easy for the formations to collide in the presence of obstacles, which will gravely affect flight safety if not guarded against. It is difficult to avoid unknown environmental obstacles using traditional modeling methods. However, artificial potential field methods can address deadlock problems such as target infeasibility and cluster congestion.
To achieve the cooperation of UAV formations without collision, a deep deterministic policy gradient (DDPG)-based centralized UAV formation control method is proposed in this study, which is designed by combining the centralized communication architecture, reinforcement learning, and artificial potential field method. First, a greedy-DDPG flight control method is studied for leader UAVs, which improves collision avoidance effectiveness. Considering maneuver constraints, reward functions, action spaces, and state spaces are improved. Additionally, to shorten the training duration, the exploration strategy of DDPG is improved using the greedy scheme. This improvement mainly uses the critic network to evaluate the value of random action groups and improves greedy selection to make actions more inclined, thus achieving rapid updates regarding the critic network and accelerating the update of the overall network. Based on this, incorporated with the artificial potential field method and leader-follower consensus, a collision-free control method is designed for followers, which can ensure collision-free following cooperation.
The numerical simulation experimental results show that the improved DDPG algorithm has a 5.9% shorter training time than the original algorithm. In the same scenario, the method that we proposed perceives the same number of obstacles as the artificial potential field method. The artificial potential field method has significant fluctuations in heading angle, while the proposed method has relatively small fluctuations. The DDPG algorithm has a smoother heading angle due to a smaller number of perceived obstacles; however, the minimum distance from the obstacles is only 9.1 m. The method that we proposed here is above 17 m from the obstacles. Furthermore, Monte Carlo experimental data under different scenarios of the long aircraft show that the ability of obstacle avoidance generalization of the proposed method is improved. Moreover, experiments were applied to the proposed formation control method. Under the same scenario and control parameters, the UAV formation control method based on the proposed architecture has lower formation errors during flight, with a maximum error of no more than 10 m. However, the artificial potential field-based formation control method has a maximum formation error of over 25 m. When encountering narrow gaps, our proposed method can quickly pass through without congestion, while the artificial potential field-based formation control method appears to hover in front of obstacles, which is not conducive to flight safety. During the entire flight, this method has a greater distance from obstacles and higher safety.
Compared with the original DDPG algorithm, the improved DDPG algorithm has faster training speed and better training effect. The formation control method can realize the formation flight of unmanned aerial vehicles under unknown obstacles. Compared with the formation control method based on artificial potential field, the formation control method avoids the hovering in place before obstacles, which is of great significance to the formation flight safety of unmanned aerial vehicles.
ISMAIL A, BAGULA B A, TUYISHIMIRE E. Internet-of-things in motion: A UAV coalition model for remote sensing in smart cities [J]. Sensors, 2018, 18(7): 2184.
YANG J H, QIAN J C, GAO H W. Forest wildfire monitoring and communication UAV system based on particle swarm optimization [J]. Journal of Physics: Conference Series, 2021, 1982(1): 012068.
TORTONESI M, STEFANELLI C, BENVEGNU E, et al. Multiple-UAV coordination and communications in tactical edge networks [J]. IEEE Communications Magazine, 2012, 50(10): 48-55.
CHEN B C, TANG W B, HUANG H Y, et al. Pop-up obstacles avoidance for UAV formation based on improved artificial potential field [J]. Computer Science, 2022, 49(S1): 686-693. (in Chinese)
LI Y, HAN W, CHEN Q Y, et al. Real-time obstacle avoidance for manned/unmanned aircraft cooperative system based on improved velocity obstacle method [J]. Journal of Northwestern Polytechnical University, 2020, 38(2): 309-318. (in Chinese)
PEDRO D, MATOS-CARVALHO J P, FONSECA J M, et al. Collision avoidance on unmanned aerial vehicles using neural network pipelines and flow clustering techniques [J]. Remote Sensing, 2021, 13(13): 2643.
LIU G, TANG J, LIU C, et al. Survey of cooperative behavior modeling technology for unmanned aerial vehicles clusters [J]. Systems Engineering and Electronics, 2021, 43(8): 2221-2231. (in Chinese)
Zhang D F, Liu F. Research and development trend of path planning based on artificial potential field method [J]. Computer Engineering & Science, 2013, 35(6): 88-95. (in Chinese)
CHEN T B, ZHANG Q S, YANG X G. On LBS shortest path correction based on improved artificial fish swarm algorithm with potential field [J]. Computer Applications and Software, 2015, 32(6): 259-262. (in Chinese)
DAI J Y, WANG C S, YIN L F, et al. Hierarchical potential field algorithm of path planning for aircraft [J]. Control Theory & Applications, 2015, 32(11): 1505-1510. (in Chinese)
QIU H T, PING X L, GAO W Y, et al. Mobile robot path planning based on improved artificial potential field method [J]. Machine Design & Research, 2017, 33(4): 36-40. (in Chinese)
LI J C, FANG Y W, CHENG H Y, et al. Large-scale fixed-wing UAV swarm system control with collision avoidance and formation maneuver [J]. IEEE Systems Journal, 2023, 17(1): 744-755.
ZHANG Y Y, WEI Y, LIU H, et al. End-to-end UAV obstacle avoidance decision based on deep reinforcement learning [J]. Journal of Northwestern Polytechnical University, 2022, 40(5): 1055-1064. (in Chinese)
WANG T H, LUO Y G, LIU J X, et al. End-to-end self-driving policy based on the deep deterministic policy gradient algorithm considering the state distribution [J]. Journal of Tsinghua University (Science and Technology), 2021, 61(9): 881-888. (in Chinese)
MNIH V, KAVUKCUOGLU K, SILVER D, et al. Human-level control through deep reinforcement learning [J]. Nature, 2015, 518(7450): 529-533.
ZHANG Y Z, XU J L, YAO K J, et al. Pursuit missions for UAV swarms based on DDPG algorithm [J]. Acta Aeronautica et Astronautica Sinica, 2020, 41(10): 324000. (in Chinese)
LI B, YANG Z P, CHEN D Q, et al. Maneuvering target tracking of UAV based on MN-DDPG and transfer learning [J]. Defence Technology, 2021, 17(2): 457-466.
GAO J P, HU X Y, JIANG Z Y, Unmanned aerial vehicle track planning algorithm based on improved DDPG [J]. Computer Engineering and Applications, 2022, 58(8): 264-272. (in Chinese)
SHEVITZ D, PADEN B. Lyapunov stability theory of nonsmooth systems [J]. IEEE Transactions on Automatic Control, 1994, 39(9): 1910-1914.