Turn off MathJax
Article Contents
ZHAO Xingpeng, XU Hang, CHANG Pengfa, et al. A semisupervised radar-based human action recognition method via multidomain collaborative training[J]. Journal of Radars, in press. doi: 10.12000/JR25223
Citation: ZHAO Xingpeng, XU Hang, CHANG Pengfa, et al. A semisupervised radar-based human action recognition method via multidomain collaborative training[J]. Journal of Radars, in press. doi: 10.12000/JR25223

A Semisupervised Radar-Based Human Action Recognition Method via Multidomain Collaborative Training

DOI: 10.12000/JR25223 CSTR: 32380.14.JR25223
Funds:  The National Cryptologic Science Fund of China under Grant (2025NCSF02059), The National Natural Science Foundation of China (42174175, 62105233), The Fundamental Research Program of Shanxi Province (202203021221090)
More Information
  • Corresponding author: XU Hang, xuhang@tyut.edu.cn
  • Received Date: 2025-11-04
    Available Online: 2026-03-05
  • To address the problem of insufficient labeled data in radar-based human action recognition, a semisupervised learning method based on multidomain collaborative training is proposed herein. This method fuses action features from the slow time–range, slow time–Doppler frequency, and range–Doppler frequency domains to construct a decision-level ensemble framework. An interdomain consistency evaluation mechanism is employed to dynamically adjust the contribution of each domain in ensemble prediction. Furthermore, a stratified confidence dynamic pseudo-labeling strategy is designed to balance pseudo-label quality and utilization rate through multilevel quality assessment and dynamic threshold calibration. In addition, a feature alignment constraint mechanism is introduced, where fast principal component analysis is utilized to extract the principal components of multidomain features. This mechanism guides the deep network to learn compact feature representations and enhances model discriminability. On the through-wall human action dataset based on a random code radar, an average recognition accuracy of (93.6±1.6)% is achieved with 5% labeled data; meanwhile, the corresponding accuracy on the indoor human action dataset based on a frequency modulated continuous-wave (FMCW) radar is (91.3±1.9)%. The proposed method outperforms supervised learning methods, including the use of Bi-LSTM, LH-ViT, and MFAFN, as well as semisupervised learning methods, including the use of FixMatch, C-TGAN, MF-Match, and LW-HGR. Experimental results demonstrate that the proposed method exhibits stable performance across two different radar systems (random code and FMCW radars) and two detection scenarios (through-wall and indoor scenarios), validating its cross-system and cross-scenario adaptability. Finally, the semisupervised learning model based on multidomain collaborative training has 1.30M parameters, requires 26.16M floating-point operations, and exhibits a size of 5.01 MB, demonstrating high computational efficiency.

     

  • loading
  • [1]
    LI Xinyu, HE Yuan, and JING Xiaojun. A survey of deep learning-based human activity recognition in radar[J]. Remote Sensing, 2019, 11(9): 1068. doi: 10.3390/rs11091068.
    [2]
    丁一鹏, 厍彦龙. 穿墙雷达人体动作识别技术的研究现状与展望[J]. 电子与信息学报, 2022, 44(4): 1156–1175. doi: 10.11999/JEIT211051.

    DING Yipeng and SHE Yanlong. Research status and prospect of human movement recognition technique using through-wall radar[J]. Journal of Electronics & Information Technology, 2022, 44(4): 1156–1175. doi: 10.11999/JEIT211051.
    [3]
    JARAMILLO I E, JEONG J G, LOPEZ P R, et al. Real-time human activity recognition with IMU and encoder sensors in wearable exoskeleton robot via deep learning networks[J]. Sensors, 2022, 22(24): 9690. doi: 10.3390/s22249690.
    [4]
    WANG Zhengwei, SHE Qi, and SMOLIC A. ACTION-Net: Multipath excitation for action recognition[C]. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2021: 13214–13223. doi: 10.1109/CVPR46437.2021.01301.
    [5]
    GUAN Qiuju, YIN Xuguang, GUO Xuemei, et al. A novel infrared motion sensing system for compressive classification of physical activity[J]. IEEE Sensors Journal, 2016, 16(8): 2251–2259. doi: 10.1109/JSEN.2016.2514606.
    [6]
    杨小鹏, 高炜程, 渠晓东. 基于微多普勒角点特征与Non-Local机制的穿墙雷达人体步态异常终止行为辨识技术[J]. 雷达学报(中英文), 2024, 13(1): 68–86. doi: 10.12000/JR23181.

    YANG Xiaopeng, GAO Weicheng, and QU Xiaodong. Human anomalous gait termination recognition via through-the-wall radar based on micro-Doppler corner features and Non-Local mechanism[J]. Journal of Radars, 2024, 13(1): 68–86. doi: 10.12000/JR23181.
    [7]
    SONG Yongkun, DAI Yongpeng, JIN Tian, et al. Dual-task human activity sensing for pose reconstruction and action recognition using 4-D imaging radar[J]. IEEE Sensors Journal, 2023, 23(19): 23927–23940. doi: 10.1109/JSEN.2023.3308788.
    [8]
    AMIN M G, ZHANG Y D, AHMAD F, et al. Radar signal processing for elderly fall detection: The future for in-home monitoring[J]. IEEE Signal Processing Magazine, 2016, 33(2): 71–80. doi: 10.1109/MSP.2015.2502784.
    [9]
    金添, 李志, 戴永鹏, 等. RIS-4D生物雷达多人体定位与生命体征监测[J]. 信号处理, 2024, 40(2): 225–235. doi: 10.16798/j.issn.1003-0530.2024.02.001.

    JIN Tian, LI Zhi, DAI Yongpeng, et al. Multi-subject localization and vital-sign monitoring with RIS-4D bioradar[J]. Journal of Signal Processing, 2024, 40(2): 225–235. doi: 10.16798/j.issn.1003-0530.2024.02.001.
    [10]
    KIM Y and LING Hao. Human activity classification based on micro-Doppler signatures using a support vector machine[J]. IEEE Transactions on Geoscience and Remote Sensing, 2009, 47(5): 1328–1337. doi: 10.1109/TGRS.2009.2012849.
    [11]
    CHOWDHURY A, DAS T, RANI S, et al. Activity recognition using ultra wide band range-time scan[C]. 2020 28th European Signal Processing Conference (EUSIPCO), Amsterdam, Netherlands, 2021: 1338–1342. doi: 10.23919/Eusipco47968.2020.9287598.
    [12]
    WANG Mingyang, ZHANG Y D, and CUI Guolong. Human motion recognition exploiting radar with stacked recurrent neural network[J]. Digital Signal Processing, 2019, 87: 125–131. doi: 10.1016/j.dsp.2019.01.013.
    [13]
    SHRESTHA A, LI Haobo, LE KERNEC J, et al. Continuous human activity classification from FMCW radar with Bi-LSTM networks[J]. IEEE Sensors Journal, 2020, 20(22): 13607–13619. doi: 10.1109/JSEN.2020.3006386.
    [14]
    NGUYEN N, NGUYEN T, PHAM M, et al. Improving human activity classification based on micro-Doppler signatures separation of FMCW radar[C]. 2023 12th International Conference on Control, Automation and Information Sciences (ICCAIS), Hanoi, Vietnam, 2023: 454–459. doi: 10.1109/ICCAIS59597.2023.10382332.
    [15]
    蒋留兵, 魏光萌, 车俐. 基于卷积神经网络的雷达人体动作识别方法[J]. 计算机应用与软件, 2019, 36(11): 168–174,234. doi: 10.3969/j.issn.1000-386x.2019.11.028.

    JIANG Liubing, WEI Guangmeng, and CHE Li. Human motion recognition method by radar based on CNN[J]. Computer Applications and Software, 2019, 36(11): 168–174,234. doi: 10.3969/j.issn.1000-386x.2019.11.028.
    [16]
    ZHENG Zhijie, PAN Jun, NI Zhikang, et al. Human posture reconstruction for through-the-wall radar imaging using convolutional neural networks[J]. IEEE Geoscience and Remote Sensing Letters, 2022, 19: 3505205. doi: 10.1109/LGRS.2021.3073073.
    [17]
    渠晓东, 王文远, 孟昊宇, 等. 基于鲁棒主成分分析及YOLOv8的穿墙雷达运动人员检测方法[J]. 信号处理, 2025, 41(8): 1390–1403. doi: 10.12466/xhcl.2025.08.008.

    QU Xiaodong, WANG Wenyuan, MENG Haoyu, et al. Moving human detection method in through-the-wall radar based on robust principal component analysis and YOLOv8[J]. Journal of Signal Processing, 2025, 41(8): 1390–1403. doi: 10.12466/xhcl.2025.08.008.
    [18]
    HUAN Sha, WANG Zhaoyue, WANG Xiaoqiang, et al. A lightweight hybrid vision transformer network for radar-based human activity recognition[J]. Scientific Reports, 2023, 13(1): 17996. doi: 10.1038/s41598-023-45149-5.
    [19]
    WU Yizhuo, FIORANELLI F, and GAO Chang. RadMamba: Efficient human activity recognition through radar-based micro-Doppler-oriented Mamba state-space model[EB/OL]. https://arxiv.org/abs/2504.12039, 2025.
    [20]
    LI Xinyu, HE Yuan, FIORANELLI F, et al. Semisupervised human activity recognition with radar micro-Doppler signatures[J]. IEEE Transactions on Geoscience and Remote Sensing, 2022, 60: 5103112. doi: 10.1109/TGRS.2021.3090106.
    [21]
    LIU Li, WANG Shengyao, SONG Chenyan, et al. Radar-based human motion recognition using semisupervised triple-GAN[J]. IEEE Sensors Journal, 2023, 23(24): 30691–30702. doi: 10.1109/JSEN.2023.3327963.
    [22]
    YUN Tianhe and WANG Zhangang. MF-Match: A semi-supervised model for human action recognition[J]. Sensors, 2024, 24(15): 4940. doi: 10.3390/s24154940.
    [23]
    SADEGHI-ADL Z and AHMAD F. Semi-supervised convolutional autoencoder with attention mechanism for activity recognition[C]. 2023 31st European Signal Processing Conference, Helsinki, Finland, 2023: 785–789. doi: 10.23919/EUSIPCO58844.2023.10289719.
    [24]
    PINYOANUNTAPONG E, ALI A, JAKKALA K, et al. GaitSADA: Self-aligned domain adaptation for mmWave gait recognition[C]. 2023 IEEE 20th International Conference on Mobile Ad Hoc and Smart Systems (MASS), Toronto, Canada, 2023: 218–226. doi: 10.1109/MASS58611.2023.00034.
    [25]
    XU Hang, LI Yong, LI Yingxin, et al. Through-wall human motion recognition using random code radar sensor with multi-domain feature fusion[J]. IEEE Sensors Journal, 2022, 22(15): 15123–15132. doi: 10.1109/JSEN.2022.3183292.
    [26]
    XU Hang, LI Yong, DONG Qingran, et al. Random code radar with range-time-frequency points and improved PointConv network for through-wall human action recognition[J]. IEEE Sensors Journal, 2025, 25(8): 13719–13728. doi: 10.1109/JSEN.2025.3548121.
    [27]
    DING Chuanwei, ZHANG Li, CHEN Haoyu, et al. Sparsity-based human activity recognition with PointNet using a portable FMCW radar[J]. IEEE Internet of Things Journal, 2023, 10(11): 10024–10037. doi: 10.1109/JIOT.2023.3235808.
    [28]
    DING Wen, GUO Xuemei, and WANG Guoli. Radar-based human activity recognition using hybrid neural network model with multidomain fusion[J]. IEEE Transactions on Aerospace and Electronic Systems, 2021, 57(5): 2889–2898. doi: 10.1109/TAES.2021.3068436.
    [29]
    CAO Lin, LIANG Song, ZHAO Zongmin, et al. Human activity recognition method based on FMCW radar sensor with multi-domain feature attention fusion network[J]. Sensors, 2023, 23(11): 5100. doi: 10.3390/s23115100.
    [30]
    WU Yajie, WANG Xiang, GUO Shisheng, et al. A lightweight network with multifeature fusion for mmWave radar-based hand gesture recognition[J]. IEEE Sensors Journal, 2024, 24(12): 19553–19561. doi: 10.1109/JSEN.2024.3395638.
    [31]
    BU Yuqing, WANG Xiang, ZHANG Bo, et al. Multidomain fusion method for human head movement recognition[J]. IEEE Transactions on Instrumentation and Measurement, 2023, 72: 2504608. doi: 10.1109/TIM.2023.3238750.
    [32]
    SOHN K, BERTHELOT D, LI Chunliang, et al. FixMatch: Simplifying semi-supervised learning with consistency and confidence[C]. The 34th International Conference on Neural Information Processing Systems, Vancouver, Canada, 2020: 51.
    [33]
    BLUM A and MITCHELL T. Combining labeled and unlabeled data with co-training[C]. The Eleventh Annual Conference on Computational Learning Theory, Madison, USA, 1998: 92–100. doi: 10.1145/279943.279962.
    [34]
    XU Yi, SHANG Lei, YE Jinxing, et al. Dash: Semi-supervised learning with dynamic thresholding[C]. The 38th International Conference on Machine Learning, Virtual Event, 2021: 11525–11536.
    [35]
    ZHU Rui, LV Songlin, WANG Zikang, et al. Bi-CoG: Bi-consistency-guided self-training for vision-language models[EB/OL]. https://arxiv.org/abs/2510.20477, 2025.
    [36]
    FIORANELLI F, SHAH S A, LI H, et al. Radar signatures of human activities[DS/OL]. University of Glasgow. https://doi.org/10.5525/gla.researchdata.848, 2019.
  • 加载中

Catalog

    通讯作者: 陈斌, bchen63@163.com
    • 1. 

      沈阳化工大学材料科学与工程学院 沈阳 110142

    1. 本站搜索
    2. 百度学术搜索
    3. 万方数据库搜索
    4. CNKI搜索
    Article views(16) PDF downloads(0) Cited by()
    Proportional views
    Related

    /

    DownLoad:  Full-Size Img  PowerPoint