一种基于深度强化学习的频率捷变雷达智能频点决策方法

张嘉翔 张凯翔 梁振楠 陈新亮 刘泉华

张嘉翔, 张凯翔, 梁振楠, 等. 一种基于深度强化学习的频率捷变雷达智能频点决策方法[J]. 雷达学报(中英文), 2024, 13(1): 227–239. doi: 10.12000/JR23197
引用本文: 张嘉翔, 张凯翔, 梁振楠, 等. 一种基于深度强化学习的频率捷变雷达智能频点决策方法[J]. 雷达学报(中英文), 2024, 13(1): 227–239. doi: 10.12000/JR23197
ZHANG Jiaxiang, ZHANG Kaixiang, LIANG Zhennan, et al. An intelligent frequency decision method for a frequency agile radar based on deep reinforcement learning[J]. Journal of Radars, 2024, 13(1): 227–239. doi: 10.12000/JR23197
Citation: ZHANG Jiaxiang, ZHANG Kaixiang, LIANG Zhennan, et al. An intelligent frequency decision method for a frequency agile radar based on deep reinforcement learning[J]. Journal of Radars, 2024, 13(1): 227–239. doi: 10.12000/JR23197

一种基于深度强化学习的频率捷变雷达智能频点决策方法

DOI: 10.12000/JR23197
基金项目: 国家自然科学基金(62201048)
详细信息
    作者简介:

    张嘉翔,博士生,主要研究方向为智能干扰感知与抗干扰决策

    张凯翔,博士生,主要研究方向为分布式雷达和抗干扰

    梁振楠,博士,副研究员,硕士生导师,主要研究方向为数字阵列雷达系统和宽带雷达信号处理

    陈新亮,博士,讲师,硕士生导师,主要研究方向为目标检测跟踪和软件化雷达

    刘泉华,博士,教授,博士生导师,主要研究方向为高分辨雷达系统及信号处理

    通讯作者:

    梁振楠 liangzhennan@bit.edu.cn

  • 责任主编:全英汇 Corresponding Editor: QUAN Yinghui
  • 中图分类号: TN958

An Intelligent Frequency Decision Method for a Frequency Agile Radar Based on Deep Reinforcement Learning(in English)

Funds: The National Natural Science Foundation of China (62201048)
More Information
  • 摘要: 自卫式干扰机发射的瞄准干扰使多种基于信号处理的被动干扰抑制方法失效,对现代雷达产生了严重威胁,频率捷变作为一种主动对抗方式为对抗瞄准干扰提供了可能。针对传统随机跳频抗干扰性能不稳定、频点选取自由度有限、策略学习所需时间长等问题,该文面向频率捷变雷达,提出了一种快速自适应跳频策略学习方法。首先设计了一种频点可重复选取的频率捷变波形,为最优解提供了更多选择。在此基础上,通过利用雷达与干扰机持续对抗收集到的数据,基于深度强化学习的探索与反馈机制,不断优化频点选取策略。具体来说,通过将上一时刻雷达频点及当前时刻感知到的干扰频点作为强化学习输入,神经网络智能选取当前时刻各子脉冲频点,并根据目标检测结果以及信干噪比两方面评价抗干扰效能,从而优化策略直至最优。从提高最优策略收敛速度出发,设计的输入状态不依赖历史时间步、引入贪婪策略平衡搜索-利用机制、配合信干噪比提高奖励差异。多组仿真实验结果表明,所提方法能够收敛到最优策略且具备较高的收敛效率。

     

  • 图  1  频率捷变波形示意图

    Figure  1.  Schematic diagram of the frequency agility waveform

    图  2  MDP的随机独立性与强化学习的优化目标

    Figure  2.  The random independence of MDP and the optimization objectives of reinforcement learning

    图  3  DQN网络参数的更新过程

    Figure  3.  The network parameter update process of DQN

    图  4  全连接神经网络结构示意图

    Figure  4.  The schematic diagram of fully connected neural network structure

    图  5  脉内侦干策略

    Figure  5.  The intra-pulse interception-jamming strategy

    图  6  脉间侦干策略

    Figure  6.  The pulse-to-pulse interception-jamming strategy

    图  7  脉内侦干策略的子脉冲频点决策训练结果

    Figure  7.  The training results of sub-pulse frequency decision for the intra-pulse interception-jamming strategy

    图  8  训练用CPI数量对脉内侦干策略下对抗成功率的影响

    Figure  8.  The impact of the number of CPI used for training on the success rate of confrontation for the intra-pulse interception-jamming strategy

    图  9  雷达与干扰对抗4个PRT的策略及对抗奖励

    Figure  9.  The strategies and rewards for radar anti-jamming during four PRT periods

    图  10  雷达执行最优策略的时频图及一维距离像

    Figure  10.  The time-frequency map and the one-dimensional High-Resolution Range Profile (HRRP) for radar executing optimal strategy

    图  11  脉间侦干策略的子脉冲频点决策训练结果

    Figure  11.  The training results of sub-pulse frequency decision for the pulse-to-pulse interception-jamming strategy

    图  12  训练用CPI数量对脉间侦干策略对抗成功率的影响

    Figure  12.  The impact of the number of CPI used for training on the success rate of confrontation for the pulse-to-pulse interception-jamming strategy

    图  13  对抗3个侦干周期的雷达策略及对抗奖励

    Figure  13.  The strategies and rewards for radar anti-jamming during three interception-jamming periods

    图  1  Schematic diagram of the frequency agility waveform

    图  2  Random independence of MDP and the optimization objectives of reinforcement learning

    图  3  Network parameter update process of DQN

    图  4  Schematic diagram of fully connected neural network structure

    图  5  Intrapulse interception-jamming strategy

    图  6  Pulse-to-pulse interception-jamming strategy

    图  7  Training results of subpulse frequency decision for the intrapulse interception-jamming strategy

    图  8  Impact of the number of CPIs used for training on the success rate of confrontation for the intrapulse interception-jamming strategy

    图  9  Strategies and rewards for radar anti-jamming during four PRT periods

    图  10  Time-frequency map and one-dimensional HRRP for radar executing optimal strategy

    图  11  Training results of subpulse frequency decision for the pulse-to-pulse interception-jamming strategy

    图  12  Impact of the number of CPIs used for training on the success rate of confrontation for the pulse-to-pulse interception-jamming strategy

    图  13  Strategies and rewards for radar anti-jamming during three interception-jamming periods

    1  基于深度Q网络的雷达子脉冲频点决策

    1.   Radar sub-pulse frequency decision based on Deep Q-Network (DQN)

     Step 1:初始化:
      Step 1-1:使用随机参数$\theta $初始化估计值${\text{Q}}$网络
      Step 1-2:使用参数${\theta ^ - }{\text{=}}\theta $初始化目标值$ {{\hat {\rm Q}}} $网络
      Step 1-3:初始化经验池D
      Step 1-4:初始化干扰策略,雷达子脉冲数量及频点,折扣因
      子$\gamma $,学习率$\alpha $,贪婪因子$\varepsilon $,软间隔更新系数$\tau $等参数
     Step 2:每幕:
     Step 2-1:设置初始状态$ {s_1} = \left[ {{f_{{\mathrm{R}},0}},{f_{{\mathrm{J}},1}}} \right] $
     Step 2-2:每个时间步:
      Step 2-2-1:使用$\varepsilon $-贪婪原则依据估计值网络的输出结果选择
      各子脉冲频点$ {a_t} = {f_{{\mathrm{R}},t}} = \left[ {{f_{{\mathrm{sub}}1,t}},{f_{{\mathrm{sub}}2,t}}, \cdots ,{f_{{\mathrm{sub}}N,t}}} \right] $,即以
      $1 - \varepsilon $概率选择估计值网络输出的最佳的频点或者以$\varepsilon $概率随
      机选择频点
      Step 2-2-2:雷达发射子脉冲频率捷变波形,接收到回波后,感
      知得到下一时刻状态$ {s_{t + 1}} $并根据目标检测结果和脉压后的信
      干噪比评估当前时刻奖励${r_t}$
      Step 2-2-3:将$\left( {{s_t},{a_t},{r_t},s{}_{t + 1}} \right)$存储到经验池D中,如果经验池
      中的样本数超出预定数量,则删除早期训练样本数据,以便存
      储并使用最新样本数据
      Step 2-2-4:如果经验池D中保存数量超过起始值,则从D中选
      择批大小(batchsize)个样本作为训练集输入到估计值和目标值
      网络中,分别计算得到$ Q\left( {{s_t},{a_t};\theta } \right) $和$ y = {r_t} + \gamma \max \hat Q ( {s_{t + 1}},$
      $a_{t + 1}';{\theta ^ - } ) $,并反向梯度求导使误差函数$L\left( \theta \right) = \left[ y - Q\left( {s_t},{a_t}; \right.\right. $
      $ \left. \left.\theta \right) \right]^2 $趋近0,更新估计值网络参数$\theta $
      Step 2-2-5:每隔一定的时间步软更新目标值网络参数${\theta ^ - }$
     Step 2-3:结束该时间步
     Step 2-4:降低贪婪概率$\varepsilon $
     Step 3:结束该幕
    下载: 导出CSV

    表  1  频率捷变信号参数设置

    Table  1.   The parameter settings of frequency agile signal

    参数 数值
    子脉冲调制类型 LFM
    子脉冲个数 3
    子脉冲频点 [10 MHz, 30 MHz, 50 MHz]
    子脉冲脉宽 5 μs
    子脉冲带宽 5 MHz
    信噪比 0 dB
    下载: 导出CSV

    表  2  干扰参数设置

    Table  2.   The parameter settings of jamming

    干扰类型 参数 数值
    窄带瞄频 瞄准频点 [10 MHz, 30 MHz, 50 MHz]
    带宽 10 MHz
    干噪比 35 dB
    宽带阻塞 带宽 120 MHz
    干噪比 30 dB
    下载: 导出CSV

    表  3  DQN参数设置

    Table  3.   The parameter settings of DQN

    参数 数值
    批大小 64
    学习率 0.001
    折扣因子 0.99
    缓冲区大小 10000
    起始训练样本量 64
    贪婪因子衰减系数 0.2
    32个时间步
    目标值网络更新周期 4个时间步
    目标值网络软间隔更新系数 0.01
    隐藏层数量 2
    隐藏层神经元个数 64
    归一化系数 80
    下载: 导出CSV

    表  4  脉内侦干策略的对抗成功率(%)

    Table  4.   The success rate of confrontation for the intra-pulse interception-jamming strategy (%)

    策略PRT对抗成功率CPI对抗成功率
    随机频点9.70
    PPO949
    DQN100100
    下载: 导出CSV

    表  5  脉内侦干策略下各种雷达策略对抗1000次结果(fJ=fsub1)

    Table  5.   The results of 1000 confrontations with various radar strategies for the intra-pulse interception-jamming strategy (fJ=fsub1)

    雷达频点选择 目标检测率(%) 信干噪比(dB) 平均得分
    [1,1,1] 0 –3.00
    [1,1,2] 0 11.09 –1.12
    [1,1,3] 0 12.25 –0.96
    [1,2,2] 97.6 15.20 1.09
    [1,2,3] 81.7 12.78 0.78
    [1,3,3] 99.7 16.06 1.19
    [2,1,1] 98.3 15.35 1.12
    [2,1,3] 75.6 12.47 0.64
    [2,3,3] 97.7 15.19 1.10
    [3,1,1] 99.6 16.07 1.18
    注:综合考虑噪声随机性引起的得分波动情况,加粗项为最优策略
    下载: 导出CSV

    表  6  脉间侦干策略的对抗成功率(%)

    Table  6.   The success rate of confrontation for the pulse-to-pulse interception-jamming strategy (%)

    策略 PRT对抗成功率 CPI对抗成功率
    随机频点 0.7 0
    PPO 93.6 31
    DQN 100 100
    下载: 导出CSV

    表  7  脉间侦干策略下各种雷达策略对抗1000次的结果(fJ=1)

    Table  7.   The results of 1000 confrontations with various radar strategies for the pulse-to-pulse interception-jamming strategy (fJ=1)

    雷达频点选择 目标检测率(%) 信干噪比(dB) 平均得分
    [1,1,1] 0 –3.00
    [1,2,3] 81.3 12.74 0.76
    [2,2,2] 99.7 17.08 3.17
    [3,3,3] 100 17.58 3.22
    注:加粗项表示最优策略
    下载: 导出CSV

    1  Radar subpulse frequency decision based on Deep Q-Network (DQN)

     Step 1: Initialization:
      Step 1-1: Initialize the estimation Q-network with random parameters $\theta $.
      Step 1-2: Initialize the target $ {{\hat Q}} $-network with parameters ${\theta ^ - }{\text{ = }}\theta $.
      Step 1-3: Initialize the experience replay buffer D.
      Step 1-4: Initialize the jamming strategy, number of radar subpulses and their frequencies, discount factor $\gamma $, learning rate $\alpha $,
      exploration factor $\varepsilon $, and soft-update coefficient $\tau $.
     Step 2: For each episode:
      Step 2-1: Set the initial state $ {s_1} = \left[ {{f_{{\mathrm{R}},0}},{f_{{\mathrm{J}},1}}} \right] $.
      Step 2-2: For each time step:
       Step 2-2-1: Use the epsilon-greedy principle to select the subpulse frequencies $ {a_t} = {f_{{\mathrm{R}},t}} = \left[ {{f_{{\mathrm{sub}}1,t}},{f_{{\mathrm{sub}}2,t}}, \cdots ,{f_{{\mathrm{sub}}N,t}}} \right] $ based on the
       output of the estimation Q-network, i.e., with probability $1 - \varepsilon $, choose the best frequency from the estimation Q-network, or with
       probability $\varepsilon $, choose a frequency randomly.
       Step 2-2-2: The radar transmits the frequency agile waveform, receives the echo, and perceives the next state $ {s_{t + 1}} $. Evaluate the
       current reward ${r_t}$ based on the target detection result and the SJNR after pulse compression.
       Step 2-2-3: Store the transition $\left( {{s_t},{a_t},{r_t},s{}_{t + 1}} \right)$ in the experience replay buffer D. If the number of samples in the buffer exceeds a
       predetermined number, remove the earliest training samples to store and use the latest samples.
       Step 2-2-4: If the number of samples in the experience replay buffer D exceeds a threshold, select a batch size of samples from D as
       the training set. Input the batch into the estimation and target networks to compute $ Q\left( {{s_t},{a_t};\theta } \right) $ and $ y = {r_t} + \gamma \max \hat Q ( {s_{t + 1}},$
        $a_{t + 1}';{\theta ^ - } ) $, respectively. Perform backpropagation to minimize the loss function $ L\left( \theta \right) = {\left[ {y - Q\left( {{s_t},{a_t};\theta } \right)} \right]^2} $, and update the
       parameters $\theta $ of the estimation Q-network.
       Step 2-2-5: Soft update the target network parameters ${\theta ^ - }$ every certain number of time steps.
     Step 2-3: End the current time step.
     Step 2-4: Decrease the exploration factor $\varepsilon $.
     Step 3: End the episode.
    下载: 导出CSV

    表  1  Parameter settings of frequency agile signal

    Parameter Value
    Subpulse modulation type LFM
    Number of subpulses 3
    Subpulse frequency [10 MHz, 30 MHz, 50 MHz]
    Subpulse width 5 μs
    Subpulse bandwidth 5 MHz
    Signal-to-Noise Ratio (SNR) 0 dB
    下载: 导出CSV

    表  2  Parameter settings of jamming

    Jamming type Parameter Value
    NBAFJ Aiming frequency [10 MHz, 30 MHz, 50 MHz]
    Bandwidth 10 MHz
    Jammer-to-Noise
    Ratio (JNR)
    35 dB
    WBBJ Bandwidth 120 MHz
    Jammer-to-Noise
    Ratio (JNR)
    30 dB
    下载: 导出CSV

    表  3  Parameter settings of DQN

    Parameter Value
    Batch size 64
    Learning rate 0.001
    Discount factor 0.99
    Buffer size 10000
    Initial training sample size 64
    Epsilon decay factor 0.2
    Episode 32 time steps
    Target network update interval 4 time steps
    Target network soft update coefficient 0.01
    Number of hidden layers 2
    Number of neurons per hidden layer 64
    Normalization coefficient 80
    下载: 导出CSV

    表  4  Success rate of confrontation for the intrapulse interception-jamming strategy (%)

    Strategy PRT success rate
    of confrontation
    CPI success rate
    of confrontation
    Random frequency decision 9.7 0
    PPO 94 9
    DQN 100 100
    下载: 导出CSV

    表  5  Results of 1,000 confrontations with various radar strategies for the intrapulse interception-jamming strategy ( f J= f sub1)

    Radar frequency selection Target detection rate (%) SJNR (dB) Average score
    [1,1,1] 0 −3.00
    [1,1,2] 0 11.09 −1.12
    [1,1,3] 0 12.25 −0.96
    [1,2,2] 97.6 15.20 1.09
    [1,2,3] 81.7 12.78 0.78
    [1,3,3] 99.7 16.06 1.19
    [2,1,1] 98.3 15.35 1.12
    [2,1,3] 75.6 12.47 0.64
    [2,3,3] 97.7 15.19 1.10
    [3,1,1] 99.6 16.07 1.18
    Note: Considering the score fluctuations caused by noise randomness, the bold items indicate the optimal strategy.
    下载: 导出CSV

    表  6  Success rate of confrontation for the pulse-to-pulse interception-jamming strategy (%)

    Strategy PRT success rate
    of confrontation
    CPI success rate
    of confrontation
    Random frequency decision 0.7 0
    PPO 93.6 31
    DQN 100 100
    下载: 导出CSV

    表  7  Results of 1,000 confrontations with various radar strategies for the pulse-to-pulse interception-jamming strategy ( f J=1)

    Radar frequency selection Target detection rate (%) SJNR (dB) Average score
    [1,1,1] 0 −3.00
    [1,2,3] 81.3 12.74 0.76
    [2,2,2] 99.7 17.08 3.17
    [3,3,3] 100 17.58 3.22
    Note: The bold items indicate the optimal strategy.
    下载: 导出CSV
  • [1] 李永祯, 黄大通, 邢世其, 等. 合成孔径雷达干扰技术研究综述[J]. 雷达学报, 2020, 9(5): 753–764. doi: 10.12000/JR20087.

    LI Yongzhen, HUANG Datong, XING Shiqi, et al. A review of synthetic aperture radar jamming technique[J]. Journal of Radars, 2020, 9(5): 753–764. doi: 10.12000/JR20087.
    [2] 崔国龙, 余显祥, 魏文强, 等. 认知智能雷达抗干扰技术综述与展望[J]. 雷达学报, 2022, 11(6): 974–1002. doi: 10.12000/JR22191.

    CUI Guolong, YU Xianxiang, WEI Wenqiang, et al. An overview of antijamming methods and future works on cognitive intelligent radar[J]. Journal of Radars, 2022, 11(6): 974–1002. doi: 10.12000/JR22191.
    [3] 李康. 雷达智能抗干扰策略学习方法研究[D]. [博士论文], 西安电子科技大学, 2021. doi: 10.27389/d.cnki.gxadu.2021.003098.

    LI Kang. Research on radar intelligent antijamming strategy learning method[D]. [Ph.D. dissertation], Xidian University, 2021. doi: 10.27389/d.cnki.gxadu.2021.003098.
    [4] JIANG Wangkui, LI Yan, LIAO Mengmeng, et al. An improved LPI radar waveform recognition framework with LDC-Unet and SSR-Loss[J]. IEEE Signal Processing Letters, 2022, 29: 149–153. doi: 10.1109/LSP.2021.3130797.
    [5] GARMATYUK D S and NARAYANAN R M. ECCM capabilities of an ultrawideband bandlimited random noise imaging radar[J]. IEEE Transactions on Aerospace and Electronic Systems, 2002, 38(4): 1243–1255. doi: 10.1109/TAES.2002.1145747.
    [6] GOVONI M A, LI Hongbin, and KOSINSKI J A. Low probability of interception of an advanced noise radar waveform with linear-FM[J]. IEEE Transactions on Aerospace and Electronic Systems, 2013, 49(2): 1351–1356. doi: 10.1109/TAES.2013.6494419.
    [7] CUI Guolong, JI Hongmin, CAROTENUTO V, et al. An adaptive sequential estimation algorithm for velocity jamming suppression[J]. Signal Processing, 2017, 134: 70–75. doi: 10.1016/j.sigpro.2016.11.012.
    [8] YU K B and MURROW D J. Adaptive digital beamforming for angle estimation in jamming[J]. IEEE Transactions on Aerospace and Electronic Systems, 2001, 37(2): 508–523. doi: 10.1109/7.937465.
    [9] DAI Huanyao, WANG Xuesong, LI Yongzhen, et al. Main-lobe jamming suppression method of using spatial polarization characteristics of antennas[J]. IEEE Transactions on Aerospace and Electronic Systems, 2012, 48(3): 2167–2179. doi: 10.1109/TAES.2012.6237586.
    [10] 鲍秋香. 频率随机捷变雷达抗扫频干扰性能仿真[J]. 舰船电子对抗, 2021, 44(5): 78–81. doi: 10.16426/j.cnki.jcdzdk.2021.05.017.

    BAO Qiuxiang. Simulation of anti-sweep jamming performance of frequency random agility radar[J]. Shipboard Electronic Countermeasure, 2021, 44(5): 78–81. doi: 10.16426/j.cnki.jcdzdk.2021.05.017.
    [11] 全英汇, 方文, 沙明辉, 等. 频率捷变雷达波形对抗技术现状与展望[J]. 系统工程与电子技术, 2021, 43(11): 3126–3136. doi: 10.12305/j.issn.1001-506X.2021.11.11.

    QUAN Yinghui, FANG Wen, SHA Minghui, et al. Present situation and prospects of frequency agility radar wave form countermeasures[J]. Systems Engineering and Electronics, 2021, 43(11): 3126–3136. doi: 10.12305/j.issn.1001-506X.2021.11.11.
    [12] MINSKY M. Steps toward artificial intelligence[J]. Proceedings of the IRE, 1961, 49(1): 8–30. doi: 10.1109/JRPROC.1961.287775.
    [13] ARULKUMARAN K, DEISENROTH M P, BRUNDAGE M, et al. Deep reinforcement learning: A brief survey[J]. IEEE Signal Processing Magazine, 2017, 34(6): 26–38. doi: 10.1109/MSP.2017.2743240.
    [14] JIANG Wen, REN Yihui, and WANG Yanping. Improving anti-jamming decision-making strategies for cognitive radar via multi-agent deep reinforcement learning[J]. Digital Signal Processing, 2023, 135: 103952. doi: 10.1016/j.dsp.2023.103952.
    [15] JIANG Wen, WANG Yanping, LI Yang, et al. An intelligent anti-jamming decision-making method based on deep reinforcement learning for cognitive radar[C]. 2023 26th International Conference on Computer Supported Cooperative Work in Design (CSCWD), Rio de Janeiro, Brazil, 2023: 1662–1666. doi: 10.1109/CSCWD57460.2023.10152833.
    [16] WEI Jingjing, WEI Yinsheng, YU Lei, et al. Radar anti-jamming decision-making method based on DDPG-MADDPG algorithm[J]. Remote Sensing, 2023, 15(16): 4046. doi: 10.3390/rs15164046.
    [17] AZIZ M M, MAUD A, and HABIB A. Reinforcement learning based techniques for radar anti-jamming[C]. 2021 International Bhurban Conference on Applied Sciences and Technologies (IBCAST), Islamabad, Pakistan, 2021: 1021–1025. doi: 10.1109/IBCAST51254.2021.9393209.
    [18] LI Kang, JIU Bo, LIU Hongwei, et al. Reinforcement learning based anti-jamming frequency hopping strategies design for cognitive radar[C]. 2018 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Qingdao, China, 2018: 1–5. doi: 10.1109/ICSPCC.2018.8567751.
    [19] LI Kang, JIU Bo, and LIU Hongwei. Deep Q-network based anti-jamming strategy design for frequency agile radar[C]. 2019 International Radar Conference (RADAR), Toulon, France, 2019: 1–5. doi: 10.1109/RADAR41533.2019.171227.
    [20] LI Kang, JIU Bo, WANG Penghui, et al. Radar active antagonism through deep reinforcement learning: A way to address the challenge of mainlobe jamming[J]. Signal Processing, 2021, 186: 108130. doi: 10.1016/j.sigpro.2021.108130.
    [21] WU Qinhao, WANG Hongqiang, LI Xiang, et al. Reinforcement learning-based anti-jamming in networked UAV radar systems[J]. Applied Sciences, 2019, 9(23): 5173. doi: 10.3390/app9235173.
    [22] AK S and BRÜGGENWIRTH S. Avoiding jammers: A reinforcement learning approach[C]. 2020 IEEE International Radar Conference (RADAR), Washington, USA, 2020: 321–326. doi: 10.1109/RADAR42522.2020.9114797.
    [23] AILIYA, YI Wei, and YUAN Ye. Reinforcement learning-based joint adaptive frequency hopping and pulse-width allocation for radar anti-jamming[C]. 2020 IEEE Radar Conference (RadarConf20), Florence, Italy, 2020: 1–6. doi: 10.1109/RadarConf2043947.2020.9266402.
    [24] ZHANG Jiaxiang and ZHOU Chao. Interrupted sampling repeater jamming suppression method based on hybrid modulated radar signal[C]. 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 2019: 1–4. doi: 10.1109/ICSIDP47821.2019.9173093.
  • 加载中
图(26) / 表(16)
计量
  • 文章访问数: 
  • HTML全文浏览量: 
  • PDF下载量: 
  • 被引次数: 0
出版历程
  • 收稿日期:  2023-10-10
  • 修回日期:  2024-01-03
  • 网络出版日期:  2024-01-11
  • 刊出日期:  2024-02-28

目录

    /

    返回文章
    返回