TY - JOUR
T1 - Joint Optimization of Jamming Type Selection and Power Control for Countering Multifunction Radar Based on Deep Reinforcement Learning
AU - Pan, Zesi
AU - Li, Yunjie
AU - Wang, Shafei
AU - Li, Yan
N1 - Publisher Copyright:
© 1965-2011 IEEE.
PY - 2023/8/1
Y1 - 2023/8/1
N2 - Developing intelligent jamming methods to combat the multifunction radar (MFR) has become a vital task in electronic warfare, because the MFR can tune into different working modes according to surrounding environment information. In this article, we present a solution to the limitations of conventional jamming methods, including strong prior knowledge dependence and inaccurate selection strategies. Specifically, we study a joint optimization of jamming type selection and power control task (JO-JTSPC) for a general radar countermeasure scenario. In particular, we first model the sequential decision-making task JO-JTSPC as a Markov decision process (MDP). Subsequently, considering the differences in the designed action space, we accordingly develop two algorithms, i.e., dueling double deep Q-learning and hybrid proximal policy optimization, to solve the optimization problem. Taking into consideration the threat level of various MFR working modes and the corresponding required jamming effect, we elaborately design the reward function of MDP as a weighted summation of the mode switching factor, jamming performance factor, and jamming power factor. Further, the learned polices of these algorithms are derived based on the designed reinforcement learning elements. Extensive simulation results demonstrate that the proposed algorithms can learn highly adaptive polices in the radar countermeasure scenarios and achieve good jamming performance.
AB - Developing intelligent jamming methods to combat the multifunction radar (MFR) has become a vital task in electronic warfare, because the MFR can tune into different working modes according to surrounding environment information. In this article, we present a solution to the limitations of conventional jamming methods, including strong prior knowledge dependence and inaccurate selection strategies. Specifically, we study a joint optimization of jamming type selection and power control task (JO-JTSPC) for a general radar countermeasure scenario. In particular, we first model the sequential decision-making task JO-JTSPC as a Markov decision process (MDP). Subsequently, considering the differences in the designed action space, we accordingly develop two algorithms, i.e., dueling double deep Q-learning and hybrid proximal policy optimization, to solve the optimization problem. Taking into consideration the threat level of various MFR working modes and the corresponding required jamming effect, we elaborately design the reward function of MDP as a weighted summation of the mode switching factor, jamming performance factor, and jamming power factor. Further, the learned polices of these algorithms are derived based on the designed reinforcement learning elements. Extensive simulation results demonstrate that the proposed algorithms can learn highly adaptive polices in the radar countermeasure scenarios and achieve good jamming performance.
KW - Deep reinforcement learning
KW - jamming type
KW - multifunction radar (MFR)
KW - power control
UR - http://www.scopus.com/inward/record.url?scp=85159836581&partnerID=8YFLogxK
U2 - 10.1109/TAES.2023.3272307
DO - 10.1109/TAES.2023.3272307
M3 - Article
AN - SCOPUS:85159836581
SN - 0018-9251
VL - 59
SP - 4651
EP - 4665
JO - IEEE Transactions on Aerospace and Electronic Systems
JF - IEEE Transactions on Aerospace and Electronic Systems
IS - 4
ER -