高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

DGCN-MFW:一种面向毫米波雷达三维点云的轻量化人体动作识别网络

丁轩宇 靳标 张贞凯

丁轩宇, 靳标, 张贞凯. DGCN-MFW:一种面向毫米波雷达三维点云的轻量化人体动作识别网络[J]. 电子与信息学报. doi: 10.11999/JEIT251087
引用本文: 丁轩宇, 靳标, 张贞凯. DGCN-MFW:一种面向毫米波雷达三维点云的轻量化人体动作识别网络[J]. 电子与信息学报. doi: 10.11999/JEIT251087
DING Xuanyu, JIN Biao, ZHANG Zhenkai. DGCN-MFW: A Lightweight Human Action Recognition Network for Millimeter-Wave Radar 3D Point Clouds[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251087
Citation: DING Xuanyu, JIN Biao, ZHANG Zhenkai. DGCN-MFW: A Lightweight Human Action Recognition Network for Millimeter-Wave Radar 3D Point Clouds[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251087

DGCN-MFW:一种面向毫米波雷达三维点云的轻量化人体动作识别网络

doi: 10.11999/JEIT251087 cstr: 32379.14.JEIT251087
基金项目: 国家自然科学基金(62571220),河南省重点研发专项(241111212500),镇江市科技计划(基础研究)项目(JC2025026),2025年江苏省研究生实践创新计划(SJCX25_2502)
详细信息
    作者简介:

    丁轩宇:男,硕士生,研究方向为雷达信号处理

    靳标:男,副教授,研究方向为雷达信号处理

    张贞凯:男,教授,研究方向为雷达信号处理

    通讯作者:

    靳标 biaojin@just.edu.cn

  • 中图分类号: TN958.94

DGCN-MFW: A Lightweight Human Action Recognition Network for Millimeter-Wave Radar 3D Point Clouds

Funds: The National Natural Science Foundation of China (62571220), Key Research and Development Project of Henan Province (241111212500), Science and Technology Plan (Basic Research) Project of Zhenjiang City ( JC2025026), 2025 Jiangsu Provincial Postgraduate Practice & Innovation Program (SJCX25_2502)
  • 摘要: 毫米波雷达三维点云能精准捕捉人体动作的空间变化细节,为动作识别提供了强鲁棒性的数据源。然而,点云固有的无序性与稀疏性限制了特征提取效率,传统方法难以有效建模其局部与全局的空间依赖关系,导致识别精度受限。为解决上述问题,该文提出一种基于动态图卷积与多特征融合的轻量化动作识别网络。该网络核心包含3个模块:(1)动态图卷积模块,通过动态构建局部邻域图结构,自适应学习鲁棒的点云特征,减少动作过渡阶段的误判;(2)多尺度特征融合模块,分层聚合局部细节与全局上下文信息,增强空间表征与行为理解能力;(3)自适应帧加权模块,依据信息熵与数据可靠性为不同时序帧分配权重,聚焦关键时序片段。在公开数据集mmWave-3DPCHM-1.0上的实验表明,所提方法对TI与Vayyar数据集上的平均识别准确率分别达到98.32%与99.48%,且仅需2.06 M参数量与4.51 GFLOPS计算量,在识别精度与模型轻量化方面均优于现有主流方法。
  • 图  1  毫米波雷达三维点云的生成流程

    图  2  左前倾动作的点云数据示例

    图  3  DGCN-MFW网络的结构

    图  4  边缘卷积构建局部有向邻域图

    图  5  帧加权可视化图

    图  6  不同参数K下的性能对比

    图  7  不同加权帧数下的性能对比

    图  8  网络在TI数据集上不同学习率下的准确率和损失曲线

    图  9  网络在Vayyar数据集上不同学习率下的准确率和损失曲线

    图  10  模型在TI数据集上的混淆矩阵

    图  11  模型在Vayyar数据集上的混淆矩阵

    表  1  不同模块组合的识别准确率对比(%)

    模块组合DGCNNMSFFAFW在TI数据集上的准确率在Vayyar数据集上的准确率
    Baseline93.5994.54
    Baseline-195.3396.82
    Baseline-297.3198.57
    DGCN-MFW(本文)98.3299.48
    下载: 导出CSV

    表  2  不同网络模型在TI数据集的动作识别准确率(%)

    模型打拳跌倒左前倾左挥手开双臂右前倾右挥手静坐下蹲站立步行平均
    PointNet78.2599.1271.3673.8964.5875.6358.9465.7286.8381.2987.6578.9276.85
    PointNet++87.4298.8875.1576.6877.3174.1658.2767.8488.7687.5293.6983.4880.62
    PCT97.9099.0080.7488.5784.0885.8496.9491.4995.5495.2198.3896.6392.64
    P4Transformer98.9298.7782.4398.1385.6298.8797.8495.1999.1799.1299.5196.5695.84
    PSTNet98.5499.0378.3090.5790.7896.0996.0694.7699.2197.4099.1898.4994.98
    DGCN-MFW(本文)99.7199.4193.5798.5895.2498.8297.9499.7199.1799.3699.0799.3198.32
    下载: 导出CSV

    表  3  不同网络模型在Vayyar数据集上的动作识别准确率(%)

    模型打拳跌倒左前倾左挥手开双臂右前倾右挥手静坐下蹲站立步行平均
    PointNet89.1298.2572.1877.0566.8881.4356.7968.9190.3675.4281.6779.3079.03
    PointNet++95.6899.8069.5265.7679.8887.6470.6574.9292.6885.1795.4686.3283.24
    PCT99.34100.0082.5990.5786.7694.2699.3193.2899.9498.5298.1499.7695.31
    P4Transformer99.8599.8092.1597.4596.1797.7997.1094.2199.1799.1798.7797.8997.46
    PSTNet98.0799.1292.8697.8292.4796.7495.0795.3298.1798.1298.0297.4896.60
    DGCN-MFW(本文)99.97100.0099.8399.2698.9997.7999.4099.93100.0099.9799.0399.5699.48
    下载: 导出CSV

    表  4  不同网络模型的计算量与复杂度

    模型规模(MB)GFLOPS参数量(M)推理延迟(ms)
    PointNet8.8535.602.3032.0727
    PointNet++5.5858.701.45214.3876
    PCT10.00180.962.62182.1140
    P4Transformer10.79142.802.8066.8684
    PSTNet7.35114.141.9096.1481
    DGCN-MFW(本文)7.894.512.0615.4716
    下载: 导出CSV
  • [1] SALTI S, SCHREER O, and DI STEFANO L. Real-time 3d arm pose estimation from monocular video for enhanced HCI[C]. The 1st ACM Workshop on Vision Networks for Behavior Analysis, Vancouver, Canada, 2008: 1–8. doi: 10.1145/1461893.1461895.
    [2] 韩宗旺, 杨涵, 吴世青, 等. 时空自适应图卷积与Transformer结合的动作识别网络[J]. 电子与信息学报, 2024, 46(6): 2587–2595. doi: 10.11999/JEIT230551.

    HAN Zongwang, YANG Han, WU Shiqing, et al. Action recognition network combining spatio-temporal adaptive graph convolution and Transformer[J]. Journal of Electronics & Information Technology, 2024, 46(6): 2587–2595. doi: 10.11999/JEIT230551.
    [3] ZHANG Yushu, JI Junhao, WEN Wenying, et al. Understanding visual privacy protection: A generalized framework with an instance on facial privacy[J]. IEEE Transactions on Information Forensics and Security, 2024, 19: 5046–5059. doi: 10.1109/TIFS.2024.3389572.
    [4] 冯翔, 刘涛, 崔文卿, 等. 基于双视角时序特征融合的毫米波雷达手势数字识别研究[J]. 电子与信息学报, 2023, 45(6): 2134–2143. doi: 10.11999/JEIT220687.

    FENG Xiang, LIU Tao, CUI Wenqing, et al. Handwriting number recognition based on millimeter-wave radar with dual-view feature fusion network[J]. Journal of Electronics & Information Technology, 2023, 45(6): 2134–2143. doi: 10.11999/JEIT220687.
    [5] JIN Biao, MA Xiao, ZHANG Zhenkai, et al. Interference-robust millimeter-wave radar-based dynamic hand gesture recognition using 2-D CNN-transformer networks[J]. IEEE Internet of Things Journal, 2024, 11(2): 2741–2752. doi: 10.1109/JIOT.2023.3293092.
    [6] JIN Biao, PENG Yu, KUANG Xiaofei, et al. Robust dynamic hand gesture recognition based on millimeter wave radar using atten-TsNN[J]. IEEE Sensors Journal, 2022, 22(11): 10861–10869. doi: 10.1109/JSEN.2022.3170311.
    [7] 丁传威, 刘芷麟, 张力, 等. 基于MIMO雷达成像图序列的切向人体姿态识别方法[J]. 雷达学报(中英文), 2025, 14(1): 151–167. doi: 10.12000/JR24116.

    DING Chuanwei, LIU Zhilin, ZHANG Li, et al. Tangential human posture recognition with sequential images based on MIMO radar[J]. Journal of Radars, 2025, 14(1): 151–167. doi: 10.12000/JR24116.
    [8] 杜兰, 李逸明, 薛世鲲, 等. 结合相似度预测和阈值自动求解的开集条件下毫米波雷达点云步态识别方法[J]. 电子与信息学报, 2025, 47(6): 1850–1863. doi: 10.11999/JEIT241034.

    DU Lan, LI Yiming, XUE Shikun, et al. Millimeter-wave radar point cloud gait recognition method under open-set conditions based on similarity prediction and automatic threshold estimation[J]. Journal of Electronics & Information Technology, 2025, 47(6): 1850–1863. doi: 10.11999/JEIT241034.
    [9] SINGH A D, SANDHA S S, GARCIA L, et al. RadHAR: Human activity recognition from point clouds generated through a millimeter-wave radar[C]. The 3rd ACM Workshop on Millimeter-wave Networks and Sensing Systems, Los Cabos, Mexico, 2019: 51–56. doi: 10.1145/3349624.3356768.
    [10] YU Chengxi, XU Zhezhuang, YAN Kun, et al. Noninvasive human activity recognition using millimeter-wave radar[J]. IEEE Systems Journal, 2022, 16(2): 3036–3047. doi: 10.1109/JSYST.2022.3140546.
    [11] CHARLES R Q, SU Hao, KAICHUN M, et al. PointNet: Deep learning on point sets for 3D classification and segmentation[C]. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, USA, 2017: 77–85. doi: 10.1109/CVPR.2017.16.
    [12] QI C R, YI Li, SU Hao, et al. PointNet++: Deep hierarchical feature learning on point sets in a metric space[C]. The 31st International Conference on Neural Information Processing Systems, Long Beach, USA, 2017: 5105–5114.
    [13] LI Xing, HUANG Qian, WANG Zhijian, et al. SequentialPointNet: A strong parallelized point cloud sequence classification network for 3D action recognition[J]. arXiv preprint arXiv: 2111.08492, 2021. doi: 10.48550/arXiv.2111.08492.
    [14] FAN Hehe, YU Xin, DING Yuhang, et al. PSTNet: Point spatio-temporal convolution on point cloud sequences[C]. 9th International Conference on Learning Representations, Austria, 2021.
    [15] 余翔, 贺登辉, 杨路. 基于STF-GNN毫米波雷达点云人体动作识别方法[J/OL]. 现代雷达, https://doi.org/10.16592/j.cnki.1004-7859.2025152, 2025.

    YU Xiang, HE Denghui, and YANG Lu. Human action recognition method based on STF-GNN for millimeter-wave radar point cloud[J/OL]. Modern Radar, https://doi.org/10.16592/j.cnki.1004-7859.2025152, 2025.
    [16] FENG Runyang, GAO Yixing, MA Xueqing, et al. Mutual information-based temporal difference learning for human pose estimation in video[C]. The IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 17131–17141. doi: 10.1109/CVPR52729.2023.01643.
    [17] PACE C D, DE NUNZIO A M, DE STEFANO C, et al. Poseidon: A ViT-based architecture for multi-frame pose estimation with adaptive frame weighting and multi-scale feature fusion[J]. arXiv preprint arXiv: 2501.08446, 2025. doi: 10.48550/arXiv.2501.08446.
    [18] LIU Zhenguang, FENG Runyang, CHEN Haoming, et al. Temporal feature alignment and mutual information maximization for video-based human pose estimation[C]. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, USA, 2022: 10996–11006. doi: 10.1109/CVPR52688.2022.01073.
    [19] PENG Hanchuan, LONG Fuhui, and DING C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2005, 27(8): 1226–1238. doi: 10.1109/TPAMI.2005.159.
    [20] WU Zonghan, PAN Shirui, CHEN Fengwen, et al. A comprehensive survey on graph neural networks[J]. IEEE Transactions on Neural Networks and Learning Systems, 2021, 32(1): 4–24. doi: 10.1109/TNNLS.2020.2978386.
    [21] WANG Yue, SUN Yongbin, LIU Ziwei, et al. Dynamic graph CNN for learning on point clouds[J]. ACM Transactions on Graphics (TOG), 2019, 38(5): 146. doi: 10.1145/3326362.
    [22] HE Kaiming, ZHANG Xiangyu, REN Shaoqing, et al. Spatial pyramid pooling in deep convolutional networks for visual recognition[C]. 13th European Conference on Computer Vision -- ECCV 2014, Zurich, Switzerland, 2014: 346–361. doi: 10.1007/978-3-319-10578-9_23.
    [23] DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16x16 words: Transformers for image recognition at scale[C]. 9th International Conference on Learning Representations, Austria, 2021.
    [24] 靳标, 孙康圣, 吴昊, 等. 基于毫米波雷达三维点云的人体动作识别数据集与方法[J]. 雷达学报(中英文), 2025, 14(1): 73–89. doi: 10.12000/JR24195.

    JIN Biao, SUN Kangsheng, WU Hao, et al. 3D point cloud from millimeter-wave radar for human action recognition: Dataset and method[J]. Journal of Radars, 2025, 14(1): 73–89. doi: 10.12000/JR24195.
    [25] GUO Menghao, CAI Junxiong, LIU, Zhengning, et al. PCT: Point cloud transformer[J]. Computational Visual Media, 2021, 7(2): 187–199. doi: 10.1007/s41095-021-0229-5.
    [26] FAN Hehe, YANG Yi, and KANKANHALLI M. Point 4D transformer networks for spatio-temporal modeling in point cloud videos[C]. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, USA, 2021: 14199–14208. doi: 10.1109/CVPR46437.2021.01398.
  • 加载中
图(11) / 表(4)
计量
  • 文章访问数:  142
  • HTML全文浏览量:  47
  • PDF下载量:  14
  • 被引次数: 0
出版历程
  • 收稿日期:  2025-10-13
  • 修回日期:  2026-02-15
  • 录用日期:  2026-03-03
  • 网络出版日期:  2026-03-15

目录

    /

    返回文章
    返回