• Electronics Optics & Control
  • Vol. 30, Issue 9, 106 (2023)
YANG Yanfeia1, ZHU Yanpingb2, HU Canb2, and ZHANG Binb2
Author Affiliations
  • 1[in Chinese]
  • 2[in Chinese]
  • show less
    DOI: 10.3969/j.issn.1671-637x.2023.09.019 Cite this Article
    YANG Yanfeia, ZHU Yanpingb, HU Canb, ZHANG Binb. A Multi-UAV Collision Avoidance Decision-Making Method Based on Reinforcement Learning[J]. Electronics Optics & Control, 2023, 30(9): 106 Copy Citation Text show less

    Abstract

    With the increasingly complex low-altitude airspace environment,the probability of conflict among UAVs performing missions is increasing.Traditional reinforcement learning algorithms of SAC and DDPG suffer from slow convergence rate and unstable convergence in solving the problem of collision avoidance among multiple UAVs in limited airspace.To solve the problems,a Multi-Agent Reinforcement Learning (MARL) method based on PPO2 algorithm is proposed.Firstly,the multi-UAV flight decision-making problem is described as a Markov decision-making process.Secondly,the state space and reward function are designed to optimize the strategy by maximizing the cumulative reward,so that the overall training is more stable and converges faster.Finally,a flight simulation scene is built based on the deep learning framework TensorFlow and the reinforcement learning environment Gym,and simulation experiments are carried out.The experimental results show that the proposed method improves the success rate of collision avoidance by about 37.74 and 49.15 percent points respectively compared with that of the SAC and DDPG algorithms,which can better solve the problem of collision avoidance among multiple UAVs,and is better in terms of convergence rate and convergence stability.
    YANG Yanfeia, ZHU Yanpingb, HU Canb, ZHANG Binb. A Multi-UAV Collision Avoidance Decision-Making Method Based on Reinforcement Learning[J]. Electronics Optics & Control, 2023, 30(9): 106
    Download Citation