[1] UALATI D U, SIMAAN M A.Effectiveness of the Nash strategies in competitive multi-team target assignment problems[J]. Transaction of Aerospace and Electronic Systems, 2007, 43(1): 126-134.
[2] LI Y, DONG Y N. Weapon-target assignment based on simulated annealing and discrete particle swarm optimization in cooperative air combat[J]. Acta Aeronautic et Astronautica Sinica, 2010, 31(3): 626-631.
[3] LIU B, QIN Z, SHAO L P. Air combat decision making for coordinated multiple target attack using collective intelligence[J]. Acta Aeronautic et Astronautica Sinica, 2010, 31(7): 1727-1739.
[4] LIU B, ZHANG X P, WANG R.Air combat decision making for coordinated multiple target attack using combinatorial auction[J]. Acta Aeronautic et Astronautica Sinica, 2010, 31(7): 1434-1444.
[5] LIU X, LIU Z, HOU W S.Improved MOPSO algorithm for multi-objective programming model of weapon-target assignment[J]. Journal of Systems Engineering and Electronics, 2013, 36(2):326-330.
[9] ABBEEL P, COATES A, QUIGLEY Met al.An application of reinforcement learning to aerobatic helicopter flight[C]//Advances in Neural Information Processing Systems, 2007.doi:10.1.1.64.4458.
[12] LUCIAN B, ROBERT B, BART D S. A comprehensive survey of multi-agent reinforcement learning[J]. IEEE Transactions on SystemsManand Cybernetics-Part C:Applications and Reviews, 2008, 38(2):156-172.
[13] HU J, WLLMAN M. Nash Q-learning for general-sum stochastic games[J]. Journal of Machine Learning Research, 2003(4):1039-1069.
[14] SPECHT D F. Probabilistic neural networks[J].Neural Networks, 1990, 3(1):109-118.