1 |
尹欣繁, 章贵川, 彭先敏, 等. 军用无人机技术智能化发展及应用[J]. 国防科技, 2018, 39 (5): 30- 34.
|
|
YIN X F , ZHANG G C , PENG X M , et al. Intelligent development and application of military UAV technology[J]. National Defense Science and Technology, 2018, 39 (5): 30- 34.
|
2 |
黄长强. 未来空战过程智能化关键技术研究[J]. 航空兵器, 2019, 26 (1): 11- 19.
|
|
HUANG C Q . Research on key technology of future air combat process intelligentization[J]. Aero Weaponry, 2019, 26 (1): 11- 19.
|
3 |
周思羽, 吴文海, 张楠, 等. 自主空战机动决策方法综述[J]. 航空计算技术, 2012, 24 (1): 27- 31.
|
|
ZHOU S Y , WU W H , ZHANG N , et al. Overview of autonomous air combat maneuver decision[J]. Aeronautical Computing Technique, 2012, 24 (1): 27- 31.
|
4 |
李世豪, 丁勇, 高振龙. 基于直觉模糊博弈的无人机空战机动决策[J]. 系统工程与电子技术, 2019, 41 (5): 1063- 1070.
|
|
LI S H , DING Y , GAO Z L . UAV air combat maneuvering decision based on intuitionistic fuzzy game theory[J]. Systems Engineering and Electronics, 2019, 41 (5): 1063- 1070.
|
5 |
李世豪.复杂空战环境下基于博弈模型的无人机机动决策方法研究[D].南京:南京航空航天大学, 2019.
|
|
LI S H. Research on UAV maneuvering decision method based on game theory in complex air combat[D]. Nanjing: Nanjing University of Aeronautics and Astronautics, 2019.
|
6 |
AUSTIN F , CARBONE G , HINZ H , et al. Game theory for automated maneuvering during air-to-air combat[J]. Journal of Guidance, Control and Dynamics, 1990, 13 (6): 1143- 1149.
|
7 |
邓可, 彭宣淇, 周德云. 基于矩阵对策与遗传算法的无人机空战决策[J]. 火力与指挥控制, 2019, 44 (12): 61- 66, 71.
|
|
DENG K , PENG X Q , ZHOU D Y . Study on air combat decision method of UAV based on matrix game and genetic algorithm[J]. Fire Control & Command Control, 2019, 44 (12): 61- 66, 71.
|
8 |
孟光磊, 罗元强, 梁宵, 等. 基于动态贝叶斯网络的空战决策方法[J]. 指挥控制与仿真, 2017, 39 (3): 49- 54.
|
|
MENG G L , LUO Y Q , LIANG X , et al. Air combat decision-making method based on dynamic Bayesian network[J]. Command Control & Simulation, 2017, 39 (3): 49- 54.
|
9 |
董彦非, 申洋, 张恒喜. 空战机动决策中的影响图方法[J]. 电光与控制, 2001, 30 (1): 49- 53.
|
|
DONG Y F , SHEN Y , ZHANG H X . Influence diagram used in air combat maneuvering decision[J]. Electronics Optics & Control, 2001, 30 (1): 49- 53.
|
10 |
傅莉, 谢福怀, 孟光磊, 等. 基于滚动时域的无人机空战决策专家系统[J]. 北京航空航天大学学报, 2015, 41 (11): 1994- 1999.
|
|
FU L , XIE F H , MENG G L , et al. An UAV air-combat decision expert system based on receding horizon control[J]. Journal of Beijing University of Aeronautics and Astronautics, 2015, 41 (11): 1994- 1999.
|
11 |
付昭旺, 李战武, 强晓明, 等. 基于滚动时域控制的战斗机空战机动决策[J]. 电光与控制, 2013, 20 (3): 20- 24.
|
|
FU Z W , LI Z W , QIANG X M , et al. Tactical decision-making method based on receding horizon control for air combat[J]. Electronics Optics & Control, 2013, 20 (3): 20- 24.
|
12 |
黄长强, 赵克新, 韩邦杰, 等. 一种近似动态规划的无人机机动决策方法[J]. 电子与信息学报, 2018, 40 (10): 166- 171.
|
|
HUANG C Q , ZHAO K X , HAN B J , et al. Maneuvering decision-making method of UAV based on approximate dynamic programming[J]. Journal of Electronics & Information Technology, 2018, 40 (10): 166- 171.
|
13 |
MCGREW J S , HOW J P , WILLIAMS B , et al. Air-combat strategy using approximate dynamic programming[J]. Journal of Guidance, Control and Dynamics, 2010, 33 (5): 1641- 1654.
|
14 |
董慧芬, 王晓丽, 高庆吉. 基于离散化分布密度的无人机粗糙决策方法[J]. 系统工程与电子技术, 2019, 41 (1): 105- 111.
|
|
DONG H F , WANG X L , GAO Q J . Method on rough decision-making for UAV based on discrete distribution density[J]. Systems Engineering and Electronics, 2019, 41 (1): 105- 111.
|
15 |
SUTTON R S , BARTO A G . Reinforcement learning: an introduction[M].
|
16 |
YANG Q M , ZHANG J D , SHI G Q . Modeling of UAV path planning based on IMM under POMDP framework[J]. Journal of Systems Engineering and Electronics, 2019, 30 (3): 545- 554.
|
17 |
SONG H , LIU C C , LAWARREE J , et al. Optimal electricity supply bidding by Markov decision process[J]. IEEE Trans.on Power Systems, 2000, 15 (2): 618- 624.
|
18 |
LILLICRAP T P , HUNT J J , PRITZEL A , et al. Continuous control with deep reinforcement learning[J]. Computer Science, 2015, 8 (6): A187.
|
19 |
IOFFE S, SZEGEDY C. Batch normalization: accelerating deep network training by reducing internal covariate shift[C]//Proc.of the International Conference on Machine Learning, 2015: 448-456.
|
20 |
KINGMA D P, BA J. Adam: a method for stochastic optimization[C]//Proc.of the 3rd International Conference for Learning Representations, 2014.
|
21 |
TESAURO G . Temporal difference learning and TD-Gammon[J]. Communications of the ACM, 1995, 38 (3): 58- 68.
|
22 |
LECUN Y , BENGIO Y , HINTON G . Deep learning[J]. Nature, 2015, 521 (7553): 436- 444.
|
23 |
WATKINS C J C H , DAYAN P . Technical note: Q-learning[J]. Machine Learning, 1992, 8 (3/4): 279- 292.
|
24 |
MNIH V , KAVUKCUOGLU K , SILVER D , et al. Human-level control through deep reinforcement learning[J]. Nature, 2015, 518 (7540): 529.
|
25 |
PETERS J , SCHAAL S . Reinforcement learning of motor skills with policy gradients[J]. Neural Networks, 2008, 21 (4): 682- 697.
|
26 |
BARNDORFF-NIELSEN O E , SHEPHARD N . Non-Gaussian Ornstein-Uhlenbeck based models and some of their uses in financial economics[J]. Journal of the Royal Statistical Society, 2001, 63 (2): 167- 241.
|