基于改进MAPPO的远距离支援协同干扰方法

    A Long-Range Support Cooperative Jamming Method Based on Improved MAPPO

    • 摘要: 针对远距离支援干扰场景下组网干扰机协同干扰的问题,本文基于多智能体强化学习,将协同干扰模型描述为一个多智能体马尔可夫决策过程,并提出了一种基于改进多智能体近端策略优化(H-MAPPO)算法,通过引入多路Actor网络,将离散与连续参数进行联合优化,提升了混合动作空间训练的收敛速度。仿真结果表明,本文所设计的协同干扰方法在收敛速度、干扰效能与资源利用率上显著优于基线算法,对组网雷达有更好的干扰效果,为远距离支援协同干扰提供了高效、稳定的资源分配策略。

       

      Abstract: To address the cooperative jamming problem of networked jammers in long-range support jamming scenarios, this paper proposes a deep reinforcement learning-based jamming resource allocation method. The jamming resource allocation model is formulated as a multi-agent Markov decision process. Furthermore, an improved Multi-Agent Proximal Policy Optimization (H-MAPPO) algorithm is proposed, which introduces a multi-branch Actor network to jointly optimize discrete and continuous parameters, thereby enhancing the convergence speed of hybrid action space training. Simulation results demonstrate that the proposed cooperative jamming method significantly outperforms the baseline algorithm in terms of convergence speed, jamming effectiveness, and resource utilization efficiency, achieving superior jamming effects on networked radars and providing an efficient and stable resource allocation strategy for long-range support cooperative jamming.

       

    /

    返回文章
    返回