高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于非对称卷积的孪生网络视觉跟踪算法

蒲磊 魏振华 侯志强 冯新喜 何玉杰

蒲磊, 魏振华, 侯志强, 冯新喜, 何玉杰. 基于非对称卷积的孪生网络视觉跟踪算法[J]. 电子与信息学报, 2022, 44(8): 2957-2965. doi: 10.11999/JEIT210472
引用本文: 蒲磊, 魏振华, 侯志强, 冯新喜, 何玉杰. 基于非对称卷积的孪生网络视觉跟踪算法[J]. 电子与信息学报, 2022, 44(8): 2957-2965. doi: 10.11999/JEIT210472
PU Lei, WEI Zhenhua, HOU Zhiqiang, FENG Xinxi, HE Yujie. Siamese Network Visual Tracking Based on Asymmetric Convolution[J]. Journal of Electronics & Information Technology, 2022, 44(8): 2957-2965. doi: 10.11999/JEIT210472
Citation: PU Lei, WEI Zhenhua, HOU Zhiqiang, FENG Xinxi, HE Yujie. Siamese Network Visual Tracking Based on Asymmetric Convolution[J]. Journal of Electronics & Information Technology, 2022, 44(8): 2957-2965. doi: 10.11999/JEIT210472

基于非对称卷积的孪生网络视觉跟踪算法

doi: 10.11999/JEIT210472
基金项目: 国家自然科学基金 (62072370, 62006240)
详细信息
    作者简介:

    蒲磊:男,1991年生,博士,讲师,研究方向为计算机视觉、目标跟踪

    魏振华:男,1983年生,硕士,副教授,研究方向为电子干扰、模式识别

    侯志强:男,1973年生,博士,教授,研究方向为图像处理、计算机视觉

    冯新喜:男,1964年生,博士,教授,研究方向为信息融合、模式识别

    何玉杰:男,1987年生,博士,讲师。研究方向为目标跟踪

    通讯作者:

    蒲磊 warmstoner@163.com

  • 中图分类号: TN911.73; TP391.4

Siamese Network Visual Tracking Based on Asymmetric Convolution

Funds: The National Natural Science Foundation of China (62072370, 62006240)
  • 摘要: 针对孪生网络对旋转变化目标特征表达能力不足的问题,该文提出了基于非对称卷积的孪生网络跟踪算法。首先利用卷积核的可加性构建非对称卷积核组,可以将其应用于任意卷积核大小的已有网络结构。接着在孪生网络跟踪框架下,对AlexNet的卷积模块进行替换,并在训练和跟踪阶段对网络进行分别设计。最后在网络的末端并联地添加3个非对称卷积核,分别经过相关运算后得到3个响应图,进行加权融合后选取最大值即为目标的位置。实验结果表明,相比于SiamFC,在OTB2015数据集上精度提高了8.7%,成功率提高了4.5%。
  • 图  1  卷积核可加性示意图

    图  2  训练阶段的卷积核输出形式

    图  3  测试阶段的卷积核输出形式

    图  4  算法流程示意图

    图  5  算法定性分析对比图

    图  6  算法的距离精度曲线图和成功率曲线

    图  7  不同属性下算法的跟踪精度对比曲线

    图  8  不同属性下算法的跟踪成功率对比曲线

    表  1  训练阶段的ACSiam网络结构

    卷积层卷积核Stride通道数模板尺寸搜索尺寸
    3127×127255×255
    Conv111×11, 1×11, 11×129659×59123×123
    Pool13×329629×2961×61
    Conv25×5, 1×5, 5×1125625×2557×57
    Pool23×3225612×1228×28
    Conv33×3, 1×3, 3×1119210×1026×26
    Conv43×3, 1×3, 3×111928×824×24
    Conv53×3, 1×3, 3×1,多分支输出11286×622×22
    下载: 导出CSV

    表  2  跟踪阶段的ACSiam网络结构

    卷积层卷积核Stride通道数模板尺寸搜索尺寸
    3127×127255×255
    Conv111×1129659×59123×123
    Pool13×329629×2961×61
    Conv25×5125625×2557×57
    Pool23×3225612×1228×28
    Conv33×3119210×1026×26
    Conv43×311928×824×24
    Conv53×3, 1×3, 3×1,多分支输出11286×622×22
    下载: 导出CSV
  • [1] RAWAT W and WANG Zenghui. Deep convolutional neural networks for image classification: A comprehensive review[J]. Neural Computation, 2017, 29(9): 2352–2449. doi: 10.1162/neco_a_00990
    [2] GIRSHICK R, DONAHUE J, DARRELL T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]. Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, USA, 2014: 580–587.
    [3] LONG J, SHELHAMER E, and DARRELL T. Fully convolutional networks for semantic segmentation[C]. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, USA, 2015: 3431–3440.
    [4] SMEULDERS A W M, CHU D M, CUCCHIARA R, et al. Visual tracking: An experimental survey[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2014, 36(7): 1442–1468. doi: 10.1109/TPAMI.2013.230
    [5] BOLME D S, BEVERIDGE J R, DRAPER B A, et al. . Visual object tracking using adaptive correlation filters[C]. 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, USA, 2010: 2544–2550.
    [6] HENRIQUES J F, CASEIRO R, MARTINS P, et al. . Exploiting the circulant structure of tracking-by-detection with kernels[C]. Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 2012: 702–715.
    [7] HENRIQUES J F, CASEIRO R, MARTINS P, et al. High-speed tracking with kernelized correlation filters[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37(3): 583–596. doi: 10.1109/tpami.2014.2345390
    [8] DANELLJAN M, KHAN F S, FELSBERG M, et al. . Adaptive color attributes for real-time visual tracking[C]. Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, USA, 2014: 1090–1097.
    [9] DANELLJAN M, HÄGER G, KHAN F S, et al. . Convolutional features for correlation filter based visual tracking[C]. Proceedings of the 2015 IEEE International Conference on Computer Vision Workshop, Santiago, Chile, 2015: 621–629.
    [10] QI Yuankai, ZHANG Shengping, QIN Lei, et al. . Hedged deep tracking[C]. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, 2016: 4303–4311.
    [11] ZHANG Tianzhu, XU Changsheng, and YANG M H. Learning multi-task correlation particle filters for visual tracking[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41(2): 365–378. doi: 10.1109/TPAMI.2018.2797062
    [12] 蒲磊, 冯新喜, 侯志强, 等. 基于空间可靠性约束的鲁棒视觉跟踪算法[J]. 电子与信息学报, 2019, 41(7): 1650–1657. doi: 10.11999/JEIT180780

    PU Lei, FENG Xinxi, HOU Zhiqiang, et al. Robust visual tracking based on spatial reliability constraint[J]. Journal of Electronics &Information Technology, 2019, 41(7): 1650–1657. doi: 10.11999/JEIT180780
    [13] PU Lei, FENG Xinxi, and HOU Zhiqiang. Learning temporal regularized correlation filter tracker with spatial reliable constraint[J]. IEEE Access, 2019, 7: 81441–81450. doi: 10.1109/ACCESS.2019.2922416
    [14] LI Feng, TIAN Cheng, ZUO Wangmeng, et al. . Learning spatial-temporal regularized correlation filters for visual tracking[C]. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 4904–4913.
    [15] 侯志强, 王帅, 廖秀峰, 等. 基于样本质量估计的空间正则化自适应相关滤波视觉跟踪[J]. 电子与信息学报, 2019, 41(8): 1983–1991. doi: 10.11999/JEIT180921

    HOU Zhiqiang, WANG Shuai, LIAO Xiufeng, et al. Adaptive regularized correlation filters for visual tracking based on sample quality estimation[J]. Journal of Electronics &Information Technology, 2019, 41(8): 1983–1991. doi: 10.11999/JEIT180921
    [16] DANELLJAN M, HÄGER G, KHAN F S, et al. . Accurate scale estimation for robust visual tracking[C]. Proceedings of the British Machine Vision Conference, Nottingham, UK, 2014: 65.1–65.11.
    [17] BERTINETTO L, VALMADRE J, HENRIQUES J F, et al. Fully-convolutional Siamese networks for object tracking[C]. European Conference on Computer Vision, Amsterdam, The Netherlands, 2016: 850–865.
    [18] GUO Qing, FENG Wei, ZHOU Ce, et al. Learning dynamic Siamese network for visual object tracking[C]. Proceedings of the 2017 IEEE International Conference on Computer Vision, Venice, Italy, 2017: 1781–1789.
    [19] LI Peixia, CHEN Boyu, OUYANG Wanli, et al. . GradNet: Gradient-guided network for visual object tracking[C]. Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Korea (South), 2019: 6161–6170.
    [20] LI Bo, YAN Junjie, WU Wei, et al. High performance visual tracking with Siamese region proposal network[C]. Computer Vision and Pattern Recognition, Salt Lake City, UT, 2018: 8971–8980.
    [21] WU Yi, LIM J, and YANG M H. Object tracking benchmark[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37(9): 1834–1848. doi: 10.1109/TPAMI.2014.2388226
    [22] MA Chao, HUANG Jiabin, YANG Xiaokang, et al. . Hierarchical convolutional features for visual tracking[C]. 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 2015: 3074–3082.
    [23] BERTINETTO L, VALMADRE J, GOLODETZ S, et al. Staple: Complementary learners for real-time tracking[C]. 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, 2016: 1401–1409.
    [24] MA Chao, YANG Xiaokang, ZHANG Chongyang, et al. Long-term correlation tracking[C]. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, USA, 2015: 5388–5396.
    [25] VALMADRE J, BERTINETTO L, HENRIQUES J, et al. . End-to-end representation learning for Correlation Filter based tracking[C]. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 2017: 5000–5008.
    [26] WANG Qiang, GAO Jin, XING Junliang, et al. DCFNet: Discriminant correlation filters network for visual tracking[EB/OL].https://arxiv.org/abs/1704.04057, 2017.
    [27] ZHANG Jianming, MA Shugao, and SCLAROFF S. MEEM: Robust tracking via multiple experts using entropy minimization[C]. Proceedings of the 13th European Conference on Computer Vision, Zurich, Switzerland, 2014: 188–203.
  • 加载中
图(8) / 表(2)
计量
  • 文章访问数:  499
  • HTML全文浏览量:  279
  • PDF下载量:  76
  • 被引次数: 0
出版历程
  • 收稿日期:  2021-05-28
  • 修回日期:  2022-05-30
  • 网络出版日期:  2022-06-13
  • 刊出日期:  2022-08-17

目录

    /

    返回文章
    返回