高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于相位偏移的深度神经网络隐蔽后门攻击策略

张恒 夏雨 任燕 杜林康 张治坤

张恒, 夏雨, 任燕, 杜林康, 张治坤. 基于相位偏移的深度神经网络隐蔽后门攻击策略[J]. 电子与信息学报. doi: 10.11999/JEIT251145
引用本文: 张恒, 夏雨, 任燕, 杜林康, 张治坤. 基于相位偏移的深度神经网络隐蔽后门攻击策略[J]. 电子与信息学报. doi: 10.11999/JEIT251145
ZHANG Heng, XIA Yu, REN Yan, DU Linkang, ZHANG Zhikun. Phase Shift-Based Covert Backdoor Attack Strategy in Deep Neural Networks[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251145
Citation: ZHANG Heng, XIA Yu, REN Yan, DU Linkang, ZHANG Zhikun. Phase Shift-Based Covert Backdoor Attack Strategy in Deep Neural Networks[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT251145

基于相位偏移的深度神经网络隐蔽后门攻击策略

doi: 10.11999/JEIT251145 cstr: 32379.14.JEIT251145
基金项目: 国家自然科学基金(61873106, 62402379, 62402431, 62441618),江苏省杰出青年科学基金项目(BK20200049)
详细信息
    作者简介:

    张恒:男,教授,研究方向为信息物理系统安全、网络安全

    夏雨:男,硕士生,研究方向为机器学习安全、网络安全

    任燕:女,博士生,研究方向为机器学习安全、网络安全

    杜林康:男,助理教授,研究方向为数据隐私保护、数据溯源与确权

    张治坤:男,教授,研究方向为可信人工智能、数据安全

    通讯作者:

    张恒 zhangheng@jou.edu.cn

  • 中图分类号: TP183

Phase Shift-Based Covert Backdoor Attack Strategy in Deep Neural Networks

Funds: The National Natural Science Foundation of China (61873106, 62402379, 62402431, 62441618), The Nature Science Foundation of Jiangsu Province for Distinguished Young Scholars (BK20200049)
  • 摘要: 后门攻击对深度神经网络(DNN)构成重大安全威胁。被植入后门的模型在遭遇特定触发器输入时会诱发预设错误输出,而对干净样本仍维持基准性能。现有研究已在空间域与频域触发器设计方面展开探索,但多数方法为确保攻击成功率(ASR),而牺牲了触发器的不可感知性。该文提出一种基于相位偏移的频域后门攻击(FDPS)方法。该方法通过离散傅里叶变换(DFT)将图像映射至频域,并在选定的频率分量上施加相位扰动以嵌入触发器。具体而言,FDPS优先针对中高频相位分量进行精细调控,以最小化幅度谱变化并避免引入可察觉的伪影。鉴于相位信息主导正弦波的相对位移,此类扰动可自然协调视觉语义,从而显著提升隐蔽性。相较于传统幅度扰动策略,相位偏移在保留图像全局结构的同时,更有效地规避了基于图像的防御检测机制。实验表明,与BadNets、Blend、WaNet及Ftrojan等基准后门攻击相比,FDPS在攻击成功率、干净样本准确率以及结构相似性指数(SSIM)等指标上均表现优越。此外,在GTSRB数据集上,仅需毒化2%的训练样本即可实现99%的攻击成功率,显著降低了攻击的样本需求与技术门槛,展现出对多样化攻击场景的更强鲁棒性与适配能力。
  • 图  1  不同攻击方式下,原图与毒化图像的空间域残差对比

    图  2  基于相位偏移的后门攻击框架

    图  3  FDPS频域单点触发机理的可视化

    图  4  同一样本在不同后门攻击下的Grad-CAM可视化(FDPS仍聚焦于目标本体)

    图  5  FDPS面对ANP时攻击成功率(ASR)与良性样本准确率(BA)随剪枝强度变化曲线(ASR下降滞后于BA)

    图  6  不同攻击方法的异常指数分布,虚线表示阈值为2

    图  7  FDPS 与干净样本在 STRIP 检测下的输出熵分布对比(分布近乎重合)

    图  8  不同注入率下,FDPS方法的攻击成功率

    表  1  不同基线方法在三个数据集上的攻击成功率(ASR)和良性样本准确率(BA)对比

    模型方法CIFAR10GTSRBCINIC10
    BA(%)ASR(%)BA(%)ASR(%)BA(%)ASR(%)
    ResNet18Clean95.00-99.20-88.02-
    BadNets94.0299.4299.1399.5285.7799.26
    Blend94.2699.8799.0299.7785.8299.89
    WaNet94.2299.7499.1099.7384.8789.61
    Ftrojan94.1099.2699.0198.9085.9099.70
    DUBA94.5599.6899.1799.9287.8599.24
    FDPS94.0299.3399.1599.1286.1199.13
    RepVGGClean91.20-99.40-87.24-
    BadNets91.0799.1099.3499.6786.7899.19
    Blend91.1098.7499.1399.3685.7698.02
    WaNet91.0797.9599.1399.3886.1688.76
    Ftrojan90.0899.1499.2999.2685.8297.84
    DUBA91.1899.7899.3499.7887.0199.02
    FDPS91.0099.1999.3099.3586.2299.11
    下载: 导出CSV

    表  2  五种后门攻击方法在三数据集上的SSIM与LPIPS对比

    数据集BadNetsBlendWaNetFtrojanDUBAFDPS
    SSIMLPIPSSSIMLPIPSSSIMLPIPSSSIMLPIPSSSIMLPIPSSSIMLPIPS
    CIFAR100.9720.00930.8640.13190.9530.00990.9540.01250.9770.00810.9720.0058
    GTSRB0.9680.03470.8790.15690.9500.05280.9120.05980.9720.02010.9750.0112
    CINIC100.9780.00650.8870.18320.9410.09510.9710.07540.9680.00720.9720.0052
    下载: 导出CSV

    表  3  不同频率位置下良性样本准确率(BA)与攻击成功率(ASR)的比较

    频率 CIFAR10 GTSSRB CINIC10
    BA(%) ASR(%) BA(%) ASR(%) BA(%) ASR(%)
    ($ u=\dfrac{1}{2}M,v=\dfrac{1}{2}N $) 94.02 99.33 99.15 99.12 86.11 99.13
    ($ u=\dfrac{3}{4}M,v=\dfrac{3}{4}N $) 93.55 98.37 99.10 98.75 85.97 97.94
    ($ u=M,v=N $) 93.11 94.18 94.18 95.47 85.49 87.55
    下载: 导出CSV
  • [1] CHENG Jian, LONG Kaifang, ZHANG Shuang, et al. Text-image scene graph fusion for multimodal named entity recognition[J]. IEEE Transactions on Artificial Intelligence, 2024, 5(6): 2828–2839. doi: 10.1109/TAI.2023.3326416.
    [2] 任海玉, 刘建平, 王健, 等. 基于大语言模型的智能问答系统研究综述[J]. 计算机工程与应用, 2025, 61(7): 1–24. doi: 10.3778/j.issn.1002-8331.2409-0300.

    REN Haiyu, LIU Jianping, WANG Jian, et al. Research on intelligent question answering system based on large language model[J]. Computer Engineering and Applications, 2025, 61(7): 1–24. doi: 10.3778/j.issn.1002-8331.2409-0300.
    [3] NOWROOZI E, JADALLA N, GHELICHKHANI S, et al. Mitigating label flipping attacks in malicious URL detectors using ensemble trees[J]. IEEE Transactions on Network and Service Management, 2024, 21(6): 6875–6884. doi: 10.1109/TNSM.2024.3447411.
    [4] 任利强, 贾舒宜, 王海鹏, 等. 基于深度学习的时间序列分类研究综述[J]. 电子与信息学报, 2024, 46(8): 3094–3116. doi: 10.11999/JEIT231222.

    REN Liqiang, JIA Shuyi, WANG Haipeng, et al. A review of research on time series classification based on deep learning[J]. Journal of Electronics & Information Technology, 2024, 46(8): 3094–3116. doi: 10.11999/JEIT231222.
    [5] 汪旭童, 尹捷, 刘潮歌, 等. 神经网络后门攻击与防御综述[J]. 计算机学报, 2024, 47(8): 1713–1743. doi: 10.11897/SP.J.1016.2024.01713.

    WANG Xutong, YIN Jie, LIU Chaoge, et al. A survey of backdoor attacks and defenses on neural networks[J]. Chinese Journal of Computers, 2024, 47(8): 1713–1743. doi: 10.11897/SP.J.1016.2024.01713.
    [6] FENG Jun, LAI Yuzhe, SUN Hong, et al. SADBA: Self-adaptive distributed backdoor attack against federated learning[C]. Proceedings of the 39th AAAI Conference on Artificial Intelligence, Philadelphia, USA, 2025: 16568–16576. doi: 10.1609/aaai.v39i16.33820.
    [7] FENG Jun, YANG L T, ZHU Qing, et al. Privacy-preserving tensor decomposition over encrypted data in a federated cloud environment[J]. IEEE Transactions on Dependable and Secure Computing, 2020, 17(4): 857–868. doi: 10.1109/TDSC.2018.2881452.
    [8] ZHANG Pengfei, SUN Hong, ZHANG Zhikun, et al. Privacy-preserving recommendations with mixture model-based matrix factorization under local differential privacy[J]. IEEE Transactions on Industrial Informatics, 2025, 21(7): 5451–5459. doi: 10.1109/tii.2025.3555993.
    [9] 杜巍, 刘功申. 深度学习中的后门攻击综述[J]. 信息安全学报, 2022, 7(3): 1–16. doi: 10.19363/J.cnki.cn10-1380/tn.2022.05.01.

    DU Wei and LIU Gongshen. A survey of backdoor attack in deep learning[J]. Journal of Cyber Security, 2022, 7(3): 1–16. doi: 10.19363/J.cnki.cn10-1380/tn.2022.05.01.
    [10] GU Tianyu, LIU Kang, DOLAN-GAVITT B, et al. BadNets: Evaluating backdooring attacks on deep neural networks[J]. IEEE Access, 2019, 7: 47230–47244. doi: 10.1109/ACCESS.2019.2909068.
    [11] SELVARAJU R R, COGSWELL M, DAS A, et al. Grad-CAM: Visual explanations from deep networks via gradient-based localization[C]. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 2017: 618–626. doi: 10.1109/ICCV.2017.74.
    [12] NGUYEN T A and TRAN A T. WaNet-imperceptible warping-based backdoor attack[C]. 9th International Conference on Learning Representations, Austria, 2021. (查阅网上资料, 未找到本条文献出版城市信息, 请确认并补充).
    [13] LIN Junyu, XU Lei, LIU Yingqi, et al. Composite backdoor attack for deep neural network by mixing existing benign features[C]. Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, USA, 2020: 113–131. doi: 10.1145/3372297.3423362. (查阅网上资料,未找到本条文献出版城市信息,请确认并补充).
    [14] WANG Tong, YAO Yuan, XU Feng, et al. An invisible black-box backdoor attack through frequency domain[C]. Proceedings of the 17th European Conference on Computer Vision, Tel Aviv, Israel, 2022: 396–413. doi: 10.1007/978-3-031-19778-9_23.
    [15] FENG Yu, MA Benteng, ZHANG Jing, et al. FIBA: Frequency-injection based backdoor attack in medical image analysis[C]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, USA, 2022: 20844–20853. doi: 10.1109/CVPR52688.2022.02021.
    [16] GAO Yansong, XU Change, WANG Derui, et al. STRIP: A defence against trojan attacks on deep neural networks[C]. Proceedings of the 35th Annual Computer Security Applications Conference, San Juan, USA, 2019: 113–125. doi: 10.1145/3359789.3359790.
    [17] XU Honghui, FANG Chuangjie, WANG Renfang, et al. Dual-enhanced high-order self-learning tensor singular value decomposition for robust principal component analysis[J]. IEEE Transactions on Artificial Intelligence, 2024, 5(7): 3564–3578. doi: 10.1109/TAI.2024.3373388.
    [18] YAN Qingsen, FENG Yixu, ZHANG Cheng, et al. HVI: A new color space for low-light image enhancement[C]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2025: 5678–5687. doi: 10.1109/CVPR52734.2025.00533.
    [19] HE Kaiming, ZHANG Xiangyu, REN Shaoqing, et al. Deep residual learning for image recognition[C]. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, 2016: 770–778. doi: 10.1109/CVPR.2016.90.
    [20] DING Xiaohan, ZHANG Xiangyu, MA Ningning, et al. RepVGG: Making VGG-style ConvNets great again[C]. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2021: 13728–13737. doi: 10.1109/CVPR46437.2021.01352.
    [21] KRIZHEVSKY A. Learning multiple layers of features from tiny images[R]. Technical Report TR-2009, 2009.
    [22] STALLKAMP J, SCHLIPSING M, SALMEN J, et al. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition[J]. Neural Networks, 2012, 32: 323–332. doi: 10.1016/j.neunet.2012.02.016.
    [23] DARLOW L N, CROWLEY E J, ANTONIOU A, et al. CINIC-10 is not ImageNet or CIFAR-10[J]. arXiv preprint arXiv: 1810.03505, 2018. doi: 10.48550/arXiv.1810.03505. (查阅网上资料,请核对文献类型及格式是否正确).
    [24] GAO Yudong, CHEN Honglong, SUN Peng, et al. A dual stealthy backdoor: From both spatial and frequency perspectives[C]. Proceedings of the 38th AAAI Conference on Artificial Intelligence, Vancouver, Canada, 2024: 1851–1859. doi: 10.1609/aaai.v38i3.27954.
    [25] WU Dongxian and WANG Yisen. Adversarial neuron pruning purifies backdoored deep models[C]. Proceedings of the 35th International Conference on Neural Information Processing Systems, 2021: 1293. (查阅网上资料, 未找到本条文献出版地信息, 请确认并补充).
    [26] WANG Bolun, YAO Yuanshun, SHAN S, et al. Neural cleanse: Identifying and mitigating backdoor attacks in neural networks[C]. Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, USA, 2019: 707–723. doi: 10.1109/SP.2019.00031.
  • 加载中
图(8) / 表(3)
计量
  • 文章访问数:  20
  • HTML全文浏览量:  5
  • PDF下载量:  5
  • 被引次数: 0
出版历程
  • 收稿日期:  2025-11-01
  • 修回日期:  2026-03-09
  • 录用日期:  2026-03-09
  • 网络出版日期:  2026-03-18

目录

    /

    返回文章
    返回