Systems Engineering and Electronics ›› 2025, Vol. 47 ›› Issue (4): 1108-1114.doi: 10.12305/j.issn.1001-506X.2025.04.07

• Sensors and Signal Processing • Previous Articles     Next Articles

Optimization of radar collaborative anti-jamming strategies based on hierarchical multi-agent reinforcement learning

Ziyi WANG, Xiongjun FU, Jian DONG, Cheng FENG   

  1. School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing 100081, China
  • Received:2024-01-19 Online:2025-04-25 Published:2025-05-28
  • Contact: Xiongjun FU

Abstract:

The sparsity of rewards in the decision-making process of radar collaborative anti-jamming makes it difficult for reinforcement learning algorithms to converge and for collaborative training. To address this issue, a hierarchical multi-agent deep deterministic policy gradient (H-MADDPG) algorithm is proposed. By accumulating sparse rewards, the convergence performance of the training process is improved, and the Harvard structure idea is introduced to separately store the training experiences of multi-agent to eliminate the confusion in experience replay. In the simulations of two and four radars network simulation, under certain strong jamming conditions, the radar detection success rate is respectively increased by 15% and 30% compared to the multi-agent deep deterministic policy gradient(MADDPG) algorithm.

Key words: radar anti-jamming, hierarchical reinforcement learning, multi-agent system, deep deterministic policy gradient (DDPG), sparse reward

CLC Number: 

[an error occurred while processing this directive]