咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Autonomous maneuver decision-m... 收藏

Autonomous maneuver decision-making for a UCAV in short-range aerial combat based on an MS-DDQN algorithm

Autonomous maneuver decision-making for a UCAV in short-range aerial combat based on an MS-DDQN algorithm

作     者:Yong-feng Li Jing-ping Shi Wei Jiang Wei-guo Zhang Yong-xi Lyu Yong-feng Li;Jing-ping Shi;Wei Jiang;Wei-guo Zhang;Yong-xi Lyu

作者机构:School of AutomationNorthwestern Polytechnical UniversityXi'an 710129China Shaanxi Province Key Laboratory of Flight Control and Simulation TechnologyXi'an 710129China 

出 版 物:《Defence Technology(防务技术)》 (Defence Technology)

年 卷 期:2022年第18卷第9期

页      面:1697-1714页

核心收录:

学科分类:0710[理学-生物学] 1101[军事学-军事思想及军事历史] 08[工学] 081105[工学-导航、制导与控制] 082503[工学-航空宇航制造工程] 0825[工学-航空宇航科学与技术] 0811[工学-控制科学与工程] 0812[工学-计算机科学与技术(可授工学、理学学位)] 

基  金:supported by the National Natural Science Foundation of China (No. 61573286) the Aeronautical Science Foundation of China (No. 20180753006) the Fundamental Research Funds for the Central Universities (3102019ZDHKY07) the Natural Science Foundation of Shaanxi Province (2019JM-163, 2020JQ-218) the Shaanxi Province Key Laboratory of Flight Control and Simulation Technology 

主  题:Unmanned combat aerial vehicle Aerial combat decision Multi-step double deep Q-network Six-degree-of-freedom Aerial combat maneuver library 

摘      要:To solve the problem of realizing autonomous aerial combat decision-making for unmanned combat aerial vehicles(UCAVs) rapidly and accurately in an uncertain environment, this paper proposes a decision-making method based on an improved deep reinforcement learning(DRL) algorithm: the multistep double deep Q-network(MS-DDQN) algorithm. First, a six-degree-of-freedom UCAV model based on an aircraft control system is established on a simulation platform, and the situation assessment functions of the UCAV and its target are established by considering their angles, altitudes, environments, missile attack performances, and UCAV performance. By controlling the flight path angle, roll angle, and flight velocity, 27 common basic actions are designed. On this basis, aiming to overcome the defects of traditional DRL in terms of training speed and convergence speed, the improved MS-DDQN method is introduced to incorporate the final return value into the previous steps. Finally, the pre-training learning model is used as the starting point for the second learning model to simulate the UCAV aerial combat decision-making process based on the basic training method, which helps to shorten the training time and improve the learning efficiency. The improved DRL algorithm significantly accelerates the training speed and estimates the target value more accurately during training, and it can be applied to aerial combat decision-making.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分