1 |
HAYKIN S . Cognitive radar-a way of the future[J]. IEEE Signal Processing Magazine, 2006, 23 (1): 30- 40.
doi: 10.1109/MSP.2006.1593335
|
2 |
BACHMANN D J , EVANS R J , MORAN B . Game theoretic analysis of adaptive radar jamming[J]. IEEE Trans.on Aerospace and Electronic Systems, 2011, 47 (2): 1081- 1100.
doi: 10.1109/TAES.2011.5751244
|
3 |
WANG B , WANG J K , SONG X , et al. Research on model and algorithm of waveform selection in cognitive radar[J]. Journal of Networks, 2010, 5 (9): 1041- 1046.
|
4 |
李云杰, 朱云鹏, 高梅国. 基于Q-学习算法的认知雷达对抗过程设计[J]. 北京理工大学学报, 2015, 35 (11): 1194- 1199.
|
|
LI Y J , ZHU Y P , GAO M G . Design of cognitive radar jamming based on Q-learning algorithm[J]. Transactions of Beijing Institute of Technology, 2015, 35 (11): 1194- 1199.
|
5 |
邢强, 贾鑫, 朱卫纲. 基于Q-学习的智能雷达对抗[J]. 系统工程与电子技术, 2018, 40 (5): 1031- 1035.
|
|
XING Q , JIA X , ZHU W G . Intelligent radar countermeasure based on Q-learning[J]. Systems Engineering and Electronics, 2018, 40 (5): 1031- 1035.
|
6 |
张柏开, 朱卫纲. 对多功能雷达的DQN认知干扰决策方法[J]. 系统工程与电子技术, 2020, 42 (4): 819- 825.
|
|
ZHANG B K , ZHU W G . DQN based decision-making method of cognitive jamming against multifunctional radar[J]. Systems Engineering and Electronics, 2020, 42 (4): 819- 825.
|
7 |
周脉成. 基于博弈论的雷达干扰决策技术研究[D]. 西安: 西安电子科技大学, 2014.
|
|
ZHOU M C. Research on radar jamming decision technology based on game theory[D]. Xi'an: Xidian University, 2014.
|
8 |
孙宏伟, 童宁宁, 孙富君. 基于D-S证据理论的电子干扰模式选择[J]. 弹箭与制导学报, 2003, (S2): 218- 220.
|
|
SUN H W , TONG N N , SUN F J . Jamming design selection based on D-S theory[J]. Journal of Projectiles Rockets Missiles and Guidance, 2003, (S2): 218- 220.
|
9 |
张思齐. 基于部分可观测马尔可夫决策过程的干扰决策研究[D]. 西安: 西安电子科技大学, 2019.
|
|
ZHANG S Q. Research on interference decision based on partially observable Markov decision process[D]. Xi'an: Xidian University, 2019.
|
10 |
NGO A V , LEE S G , CHUNG T C . Bayes-adaptive hierarchical MDPs[J]. Applied Intelligence, 2016, 45 (1): 112- 126.
doi: 10.1007/s10489-015-0742-2
|
11 |
RICHARD D , EDWARD J . The optimal control of partially observable Markov processes over a finite horizon[J]. Operations Research, 1973, 21 (5): 1019- 1175.
doi: 10.1287/opre.21.5.1019
|
12 |
RAZIEH G , HOSSEIN A M , FALLAHNEZ M . A POMDP framework to find optimal policy in sustainable maintenance[J]. Scientia Iranica, 2020, 27 (3): 1544- 1561.
|
13 |
孟磊, 吴芝亮, 王轶强. POMDP模型在多机器人环境探测中的应用研究[J]. 机械科学与技术, 2022, 41 (2): 178- 185.
|
|
MENG L , WU Z L , WANG Y Q . Research on multi-robot environment exploration using POMDP[J]. Mechanical Science and Technology for Aerospace Engineering, 2022, 41 (2): 178- 185.
|
14 |
XIANG X C , SIMON F . Recent advances in deep reinforcement learning applications for solving partially observable markov decision processes (POMDP) problems[J]. Machine Learning and Knowledge Extraction, 2021, 3 (3): 554- 581.
|
15 |
WANG T X, TAGHVAEI A, MEHTA P G. Q-learning for POMDP: an application to learning locomotion gaits[C]//Proc. of the IEEE 58th Conference on Decision and Control, 2019: 2758-2763.
|
16 |
FU Y H, LIANG X X, HUANG M K, et al. Coordinating multi-agent deep reinforce-ment learning in wargame[C]//Proc. of the 3rd International Conference on Algorithms, Computing and Artificial Intelligence, 2020: 38-42.
|
17 |
SOONHO H , SIN G . Design of control framework based on deep reinforcement learning and Monte-Carlo sampling in downstream separation[J]. Computers & Chemical Engineering, 2020, 140 (2): 106910.
|
18 |
PARK H , SIM M K , CHOI D G . An intelligent financial portfolio trading strategy using deep Q-learning[J]. Expert Systems with Applications, 2020, 158 (15): 113573.
|