高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

地空跨视角定位研究综述

胡迪 袁夏 徐孝强 赵春霞

胡迪, 袁夏, 徐孝强, 赵春霞. 地空跨视角定位研究综述[J]. 电子与信息学报. doi: 10.11999/JEIT250167
引用本文: 胡迪, 袁夏, 徐孝强, 赵春霞. 地空跨视角定位研究综述[J]. 电子与信息学报. doi: 10.11999/JEIT250167
HU Di, YUAN Xia, XU Xiaoqiang, ZHAO Chunxia. A Review of Ground-to-Aerial Cross-View Localization Research[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT250167
Citation: HU Di, YUAN Xia, XU Xiaoqiang, ZHAO Chunxia. A Review of Ground-to-Aerial Cross-View Localization Research[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT250167

地空跨视角定位研究综述

doi: 10.11999/JEIT250167 cstr: 32379.14.JEIT250167
详细信息
    作者简介:

    胡迪:男,博士生,研究方向为跨视角定位、智能机器人

    袁夏:男,副教授,研究方向为计算机视觉、机器人导航

    徐孝强:男,硕士生,研究方向为视觉环境感知、智能机器人

    赵春霞:女,教授,研究方向为模式识别、图像处理、人工智能

    通讯作者:

    袁夏 yuanxia@njust.edu.cn

  • 中图分类号: TP751.1

A Review of Ground-to-Aerial Cross-View Localization Research

  • 摘要: 地空跨视角定位指的是利用空中视角图像作为参考,确定待查询的地面传感器位姿的过程。随着无人驾驶汽车、无人机导航等技术的发展,地空跨视角定位技术在智能交通、城市管理等领域扮演着越来越重要的角色。地空跨视角定位技术的发展经历了几个阶段:从早期基于人工设计特征的传统方法,到深度学习技术的引入衍生出的度量学习、图像变换及图像生成方法,以及结合距离传感器的定位技术。这些技术的发展标志着从依赖手工特征到自动学习特征表示的转变,以及从单一模态到多模态数据融合的进步。尽管取得了丰富的成果,地空跨视角定位领域仍存在很多挑战,尤其是在处理时空差异导致的定位误差问题上。这些挑战包括季节变化、日夜交替、天气状况差异等时间维度上的差异,以及视角改变、场景布局变化等空间维度上的差异,这些都对算法的鲁棒性和准确性提出了更高的要求。本文聚焦于地空跨视角定位,系统梳理了主要方法、关键数据集和评价方法,并对未来的发展趋势进行了分析。同时,本文首次系统化地整理了结合距离传感器的地空跨视角定位算法,为该领域的研究提供了新的视角和思路。
  • 图  1  地空跨视角定位方法分类

    图  2  地空跨视角定位任务图示

    图  3  地空跨视角定位方法时间轴概述

    图  4  SAFA[35]网络框架。算法首先将卫星图进行极坐标变换,将输出的转换卫星图与地面周视图进行特征匹配。

    图  5  S2SP[40]网络框架,根据高度分布完成卫星图到街景图的几何投影,并通过生成器构建地面周视图,与真实地面周视图进行匹配。

    图  6  CVML[41]网络框架。算法通过地面与卫星图像描述子间的相似度作为地空图像融合的一个特征,实现细粒度度量定位。

    图  7  面向道路结构的跨视角定位方法架构[18],分别提取地面与卫星视角下的路缘结构,并将其作为地空共享语义进行匹配并生成位姿。

    表  1  位置检索四种类别方法对比

    维度人工特征方法度量学习方法图像变换方法图像生成方法
    核心目的基于设计特征进行关联匹配学习判别性特征进行匹配显式建模补偿视角几何差异生成目标视图图像
    主要驱动力特征工程、先验模型深度表征学习、优化目标几何投影模型生成对抗网络
    核心价值奠基性、可解释性高精度、强大特征表示能力增强几何鲁棒性有效弥合视角域差异
    核心创新特定特征设计与利用自动特征学习机制几何变换规则视角映射与图像合成
    主要优势计算要求低、设计目标精准精度高、泛化性强几何鲁棒性强缩小域差距、无监督潜力大
    主要局限鲁棒性差、表达能力瓶颈算力依赖高、几何建模隐式简化现实几何变换畸变风险生成质量依赖高、模型复杂
    演进趋势逐渐被深度学习模型替代结合Transformer持续发展解决特定几何挑战探索域差异新解法
    下载: 导出CSV

    表  2  跨视角定位常用数据集

    名称介绍特点视觉数据
    (空-地)
    有无
    GNSS
    有无
    IMU
    任务目标
    CVUSA[22]涵盖纽约、洛杉矶、迈阿密等美国20个城市,包含35532对用于训练的图像和8884对用于测试的图像。支持多尺度的跨视角训练,通过引入同一地理位置的多尺度航拍图,提高了模型在
    大规模应用场景中的定位精度。
    图像-图像检索定位
    CVACT[50]覆盖了澳大利亚首都堪培拉的300平方英里的地理区域,具体包括了35532组训练用图像对和92802组测试用图像对。通过谷歌地图API采集地空图像对,并对每个街景图配备GNSS标签。用于验证方法在更大地理范围内的跨视角定位实例上的泛化能力,并引入了一种全新的Siamese CNN架构,能够同时从外观和方向几何信息中学习特征嵌入。图像-图像检索定位
    VIGOR[21]覆盖了美国4个主要城市(旧金山、纽约、西雅图和芝加哥),收集了共计90618张航拍图像和238696张街景全景图。通过密集采样来实现区域的无缝覆盖,确保了即便街景图像出现在任意位置,至少被一张航拍
    图像覆盖。
    突破现有数据集一对一检索限制,引入了航拍图像是在查询图像出现之前捕获这一更贴近实际场景的假设,构建了一个更加现实的跨视角地理定位问题框架,并首次从米级实际距离的角度评估了定位准确性。图像-图像检索定位/
    位姿回归
    KITTI[20]在德国卡尔斯鲁厄周边不同交通场景下,通过地面车载传感器记录了长达6小时的交通数据,涵盖了公路、乡村到城市的多样场景,包含大量静态和动态物体。场景趋于城乡多元化,融合了点云数据
    作为环境感知输入
    图像-图像/
    图像-点云
    位姿回归
    Oxford
    RobotCar[51,52]
    通过一个固定的路线定期穿越中心牛津市区,平均每周两次,在19个月期间共记录了
    1010公里的驾驶路程。数据集包括2000万张图像,这些图像来源于车辆上安装的六个摄像头,以及包括LiDAR、GNSS和惯性导航系统在内的地面实况数据
    偏向于同区域内的长时跨季数据,采集期间的道路和建筑施工导致路线的某些部分从数据收集开始到结束发生了显著变化图像-图像/
    图像-点云
    位姿回归
    Oxford Radar
    RobotCar[53]
    在Oxford RobotCar原平台基础上进行了扩展,在加装毫米波雷达的前提下,历时一个月在牛津中心遍历32圈,共计约280公里。通过毫米波雷达发射出的电磁波对雨、雾、夜间等场景重新扫描,实现相同区域内生成光学数据与电磁数据,彼此互补。相较于Oxford RobotCar数据集,Oxford Radar RobotCar聚焦毫米波雷达,更偏向于鲁棒性与扫描范围,通过将雷达里程计与视觉-惯导-激光一起做全局优化后,得到的轨迹弥补了在城区漂移较大的缺陷。图像-图像/
    图像-点云
    位姿回归
    Ford Multi-
    AV[54]
    由福特自动驾驶汽车车队V1、V2、V3三辆车辆在2017到2018年的不同日期和时间收集。这些车辆在密歇根州的平均行驶里程为66公里,涵盖多种驾驶场景。包含了动态城市环境中天气、光照、建筑和交通条件的季节性变化,有助于设计的算法对季节和城市变化更具鲁棒性图像-图像/
    图像-点云
    位姿回归
    下载: 导出CSV

    表  3  CVUSA和CVACT数据集模型表现

    方法CVUSACVACT
    r@1r@5r@10r@1%r@1r@5r@10r@1%
    度量学习Workman等人[22]---34.30----
    Vo等人[14]---63.70----
    Zhai等人[15]---43.20----
    CVM-Net[32]22.4749.9863.1893.625.4114.7925.6354.53
    L2LTR[55]94.0598.2798.9999.6760.7285.8589.8896.12
    TransGeo[33]94.0898.3699.0499.77----
    OR-CVFI[58]98.0599.4899.64-91.0396.3197.02-
    图像变换GeoDTR+[59]95.4098.4499.0599.7567.5789.8492.5798.54
    Liu等人[50]40.7966.8276.3696.1219.9034.8241.2363.79
    AMPLE[60]93.2297.9098.7699.7585.6993.4294.6697.58
    ArcGeo[61]97.4799.48-99.6790.9095.8496.77-
    HADGEO[62]95.0198.4599.10-86.4894.2195.50-
    SAFA[35]89.8496.9398.1499.6455.5079.9485.0894.49
    Shi等人[56]91.9697.5098.5499.6735.5560.1767.9586.71
    图像生成Regmi等人[57]48.75-81.2795.98----
    Toker等人[26]92.5697.5598.3399.5761.2985.1389.1498.32
    下载: 导出CSV

    表  4  VIGOR数据集模型表现

    方法同区域跨区域
    r@1r@5r@10r@1%Hit rater@1r@5r@10r@1%Hit rate
    SAFA[35]18.6943.6455.3697.5521.902.778.6112.9462.643.16
    SAFA+Mining[21]38.0262.8771.1297.6341.819.2321.1228.0277.849.92
    VIGOR[21]41.0765.8174.0598.3744.7111.0023.5630.7680.2211.64
    TransGeo[33]61.4887.5491.8899.5673.0918.9938.2446.9188.9421.21
    SAIG[64]55.6081.83-99.4363.5722.3542.43-90.8324.69
    EP-BEV[63]82.1897.1098.1799.70-72.1988.6891.6898.56-
    InfoNCE[65]76.1594.4696.2899.5387.24-----
    GeoDTR[66]56.5180.3786.2199.2561.7630.0252.6761.4594.4030.19
    GeoDTR+[59]59.0181.7787.1099.0767.4136.0159.0667.2294.9539.40
    下载: 导出CSV

    表  5  KITTI数据集模型表现

    类型方法有无
    先验
    同区域跨区域
    均值误差中值误差均值误差中值误差
    距离(m)角度(°)距离(m)角度(°)距离(m)角度(°)距离(m)角度(°)
    图像到图像LM[67]12.083.7211.422.8312.583.9512.113.03
    SliceMatch[68]7.964.124.393.6513.504.209.776.61
    CCVPE[69]1.220.670.620.549.161.553.330.84
    HC-Net[70]0.800.450.500.338.473.224.571.63
    BevSplat[71]2.860.332.000.286.240.332.680.28
    Petalview[72]2.103.94------
    CVR[21]9.36-8.49-----
    CVML[41]1.40-0.99-----
    Song等人[42]1.480.490.470.307.972.173.521.21
    LM[67]×15.5189.9115.9790.7515.5089.8416.0289.85
    SliceMatch[68]×9.398.715.414.4214.8523.6411.857.96
    CCVPE[69]×6.8815.013.476.1213.9477.8410.9863.84
    Petalview[72]×2.328.86--23.7957.10--
    图像到雷达RSL-Net[44]×----3.711.59--
    Tang等人[73]×----4.371.67--
    Tang等人[47]×----4.023.70--
    Tang等人[48]×----3.763.36--
    AVBM[74]×----4.965.134.044.22
    Hu等人[18]×----3.661.852.951.62
    下载: 导出CSV
  • [1] SHI Yujiao, YU Xin, LIU Liu, et al. Optimal feature transport for cross-view image geo-localization[C]. Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, USA, 2020: 11990–11997. doi: 10.1609/aaai.v34i07.6875.
    [2] WANG Tingyu, ZHENG Zhedong, YAN Chenggang, et al. Each part matters: Local patterns facilitate cross-view geo-localization[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2022, 32(2): 867–879. doi: 10.1109/TCSVT.2021.3061265.
    [3] WANG Shan, ZHANG Yanhao, PERINCHERRY A, et al. View consistent purification for accurate cross-view localization[C]. Proceedings of 2023 IEEE/CVF International Conference on Computer Vision, Paris, France, 2023: 8163–8172. doi: 10.1109/ICCV51070.2023.00753.
    [4] RODRIGUES R and TANI M. Are these from the same place? Seeing the unseen in cross-view image geo-localization[C]. Proceedings of 2021 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, USA, 2021: 3752–3760. doi: 10.1109/WACV48630.2021.00380.
    [5] HU Di, YUAN Xia, and ZHAO Chunxia. Active layered topology mapping driven by road intersection[J]. Knowledge-Based Systems, 2025, 315: 113305. doi: 10.1016/j.knosys.2025.113305.
    [6] DURGAM A, PAHEDING S, DHIMAN V, et al. Cross-view geo-localization: A survey[J]. IEEE Access, 2024, 12: 192028–192050. doi: 10.1109/ACCESS.2024.3507280.
    [7] ZHANG Kai, YUAN Xia, CHEN Shuntong, et al. Multi-modality semantic-shared cross-view ground-to-aerial localization[C]. Proceedings of the 6th ACM International Conference on Multimedia in Asia, Auckland, New Zealand, 2024: 72. doi: 10.1145/3696409.3700233.
    [8] WILSON D, ZHANG Xiaohan, SULTANI W, et al. Image and object geo-localization[J]. International Journal of Computer Vision, 2024, 132(4): 1350–1392. doi: 10.1007/s11263-023-01942-3.
    [9] 张硝, 高艺, 夏宇翔, 等. 跨视角图像地理定位数据集综述[J]. 遥感学报, 2025, 29(8): 2511–2530. doi: 10.11834/jrs.20254348.

    ZHANG Xiao, GAO Yi, XIA Yuxiang, et al. Review of cross-view image geo-localization datasets[J]. National Remote Sensing Bulletin, 2025, 29(8): 2511–2530. doi: 10.11834/jrs.20254348.
    [10] ASPERTI A, FIORILLA S, NARDI S, et al. A review of recent techniques for person re-identification[J]. Machine Vision and Applications, 2025, 36(1): 25. doi: 10.1007/s00138-024-01622-3.
    [11] ZHANG Yuxin, GUI Jie, CONG Xiaofeng, et al. A comprehensive survey and taxonomy on point cloud registration based on deep learning[C]. Proceedings of the 33rd International Joint Conference on Artificial Intelligence, Jeju, South Korea, 2024: 8344–8353. doi: 10.24963/ijcai.2024/922.
    [12] BANSAL M, SAWHNEY H S, CHENG Hui, et al. Geo-localization of street views with aerial image databases[C]. Proceedings of the 19th ACM International Conference on Multimedia, Scottsdale, USA, 2011: 1125–1128. doi: 10.1145/2072298.2071954.
    [13] BANSAL M, DANIILIDIS K, and SAWHNEY H. Ultrawide baseline facade matching for geo-localization[M]. ZAMIR A R, HAKEEM, A VAN GOOL L, et al. Large-Scale Visual Geo-Localization. Cham: Springer, 2016: 77–98. doi: 10.1007/978-3-319-25781-5_5.
    [14] VO N N and HAYS J. Localizing and orienting street views using overhead imagery[C]. Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 2016: 494–509. doi: 10.1007/978-3-319-46448-0_30.
    [15] ZHAI Menghua, BESSINGER Z, WORKMAN S, et al. Predicting ground-level scene layout from aerial imagery[C]. Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 2017: 4132–4140. doi: 10.1109/CVPR.2017.440.
    [16] MILLER I D, COWLEY A, KONKIMALLA R, et al. Any way you look at it: Semantic crossview localization and mapping with LiDAR[J]. IEEE Robotics and Automation Letters, 2021, 6(2): 2397–2404. doi: 10.1109/LRA.2021.3061332.
    [17] KIM J and KIM J. Fusing lidar data and aerial imagery with perspective correction for precise localization in urban canyons[C]. Proceedings of 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems, Macau, China, 2019: 5298–5303. doi: 10.1109/IROS40897.2019.8967711.
    [18] HU Di, YUAN Xia, XI Huiying, et al. Road structure inspired UGV-satellite cross-view geo-localization[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2024, 17: 16767–16786. doi: 10.1109/JSTARS.2024.3457756.
    [19] SHI Qian, HE Da, LIU Zhengyu, et al. Globe230k: A benchmark dense-pixel annotation dataset for global land cover mapping[J]. Journal of Remote Sensing, 2023, 3: 0078. doi: 10.34133/remotesensing.0078.
    [20] GEIGER A, LENZ P, STILLER C, et al. Vision meets robotics: The kitti dataset[J]. The International Journal of Robotics Research, 2013, 32(11): 1231–1237. doi: 10.1177/0278364913491297.
    [21] ZHU Sijie, YANG Taojiannan, and CHEN Chen. VIGOR: Cross-view image geo-localization beyond one-to-one retrieval[C]. Proceedings of 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2021: 5316–5325. doi: 10.1109/CVPR46437.2021.00364.
    [22] WORKMAN S, SOUVENIR R, and JACOBS N. Wide-area image geolocalization with aerial reference imagery[C]. Proceedings of 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 2015: 3961–3969. doi: 10.1109/ICCV.2015.451.
    [23] 黄高爽, 周杨, 胡校飞, 等. 图像地理定位研究进展[J]. 地球信息科学学报, 2023, 25(7): 1336–1362. doi: 10.12082/dqxxkx.2023.230073.

    HUANG Gaoshuang, ZHOU Yang, HU Xiaofei, et al. A survey of the research progress in image geo-localization[J]. Journal of Geo-information Science, 2023, 25(7): 1336–1362. doi: 10.12082/dqxxkx.2023.230073.
    [24] MA Yuexin, WANG Tai, BAI Xuyang, et al. Vision-centric BEV perception: A survey[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46(12): 10978–10997. doi: 10.1109/TPAMI.2024.3449912.
    [25] 周博文, 李阳, 马鑫骥, 等. 深度学习的跨视角地理定位方法综述[J]. 中国图象图形学报, 2024, 29(12): 3543–3563. doi: 10.11834/jig.230858.

    ZHOU Bowen, LI Yang, MA Xinji, et al. A survey of cross-view geo-localization methods based on deep learning[J]. Journal of Image and Graphics, 2024, 29(12): 3543–3563. doi: 10.11834/jig.230858.
    [26] TOKER A, ZHOU Qunjie, MAXIMOV M, et al. Coming down to earth: Satellite-to-street view synthesis for geo-localization[C]. Proceedings of 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2021: 6484–6493. doi: 10.1109/CVPR46437.2021.00642.
    [27] TANG Hao, LIU Hong, and SEBE N. Unified generative adversarial networks for controllable image-to-image translation[J]. IEEE Transactions on Image Processing, 2020, 29: 8916–8929. doi: 10.1109/TIP.2020.3021789.
    [28] LI Ang, MORARIU V I, and DAVIS L S. Planar structure matching under projective uncertainty for geolocation[C]. Proceedings of the 13th European Conference on Computer Vision, Zurich, Switzerland, 2014: 265–280. doi: 10.1007/978-3-319-10584-0_18.
    [29] LIN T Y, BELONGIE S, and HAYS J. Cross-view image geolocalization[C]. Proceedings of 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, USA, 2013: 891–898. doi: 10.1109/CVPR.2013.120.
    [30] CAO Song and SNAVELY N. Graph-based discriminative learning for location recognition[C]. Proceedings of 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, USA, 2013: 700–707. doi: 10.1109/CVPR.2013.96.
    [31] WORKMAN S and JACOBS N. On the location dependence of convolutional neural network features[C]. Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, USA, 2015: 70–78. doi: 10.1109/CVPRW.2015.7301385.
    [32] HU Sixing, FENG Mengdan, NGUYEN R M H, et al. CVM-Net: Cross-view matching network for image-based ground-to-aerial geo-localization[C]. Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 7258–7267. doi: 10.1109/CVPR.2018.00758.
    [33] ZHU Sijie, SHAH M, and CHEN Chen. TransGeo: Transformer is all you need for cross-view image geo-localization[C]. Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, USA, 2022: 1152–1161. doi: 10.1109/CVPR52688.2022.00123.
    [34] MEI Shaohui, LIAN Jiawei, WANG Xiaofei, et al. A comprehensive study on the robustness of deep learning-based image classification and object detection in remote sensing: Surveying and benchmarking[J]. Journal of Remote Sensing, 2024, 4: 0219. doi: 10.34133/remotesensing.0219.
    [35] SHI Yujiao, LIU Liu, YU Xin, et al. Spatial-aware feature aggregation for cross-view image based geo-localization[C]. Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, Canada, 2019: 905. doi: 10.5555/3454287.3455192.
    [36] MI Li, XU Chang, CASTILLO-NAVARRO J, et al. ConGeo: Robust cross-view geo-localization across ground view variations[C]. Proceedings of the 18th European Conference on Computer Vision, Milan, Italy, 2024: 214–230. doi: 10.1007/978-3-031-72630-9_13.
    [37] ZHANG Xiaohan, LI Xingyu, SULTANI W, et al. Cross-view geo-localization via learning disentangled geometric layout correspondence[C]. Proceedings of the 37th AAAI Conference on Artificial Intelligence, Washington, USA, 2023: 3480–3488. doi: 10.1609/aaai.v37i3.25457.
    [38] REGMI K and BORJI A. Cross-view image synthesis using geometry-guided conditional GANs[J]. Computer Vision and Image Understanding, 2019, 187: 102788. doi: 10.1016/j.cviu.2019.07.008.
    [39] ZHAO Luying, ZHOU Yang, HU Xiaofei, et al. Street-to-satellite view synthesis for cross-view geo-localization[C]. Proceedings of SPIE 13166, International Conference on Remote Sensing Technology and Survey Mapping, Changchun, China, 2024: 1316608. doi: 10.1117/12.3029089.
    [40] SHI Yujiao, CAMPBELL D, YU Xin, et al. Geometry-guided street-view panorama synthesis from satellite imagery[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 44(12): 10009–10022. doi: 10.1109/TPAMI.2022.3140750.
    [41] XIA Zimin, BOOIJ O, MANFREDI M, et al. Visual cross-view metric localization with dense uncertainty estimates[C]. Proceedings of the 17th European Conference on Computer Vision, Tel Aviv, Israel, 2022: 90–106. doi: 10.1007/978-3-031-19842-7_6.
    [42] SONG Zhenbo, ZE Xianghui, LU Jianfeng, et al. Learning dense flow field for highly-accurate cross-view camera localization[C]. Proceedings of the 37th International Conference on Neural Information Processing Systems, New Orleans, USA, 2023: 3094. doi: 10.5555/3666122.3669216.
    [43] FERVERS F, BULLINGER S, BODENSTEINER C, et al. Uncertainty-aware vision-based metric cross-view geolocalization[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 21621–21631. doi: 10.1109/CVPR52729.2023.02071.
    [44] TANG T Y, DE MARTINI D, BARNES D, et al. RSL-Net: Localising in satellite images from a radar on the ground[J]. IEEE Robotics and Automation Letters, 2020, 5(2): 1087–1094. doi: 10.1109/LRA.2020.2965907.
    [45] FU Mengyin, ZHU Minzhao, YANG Yi, et al. LiDAR-based vehicle localization on the satellite image via a neural network[J]. Robotics and Autonomous Systems, 2020, 129: 103519. doi: 10.1016/j.robot.2020.103519.
    [46] CHEN Lei, FENG Changzhou, MA Yunpeng, et al. A review of rigid point cloud registration based on deep learning[J]. Frontiers in Neurorobotics, 2024, 17: 1281332. doi: 10.3389/fnbot.2023.1281332.
    [47] TANG T Y, DE MARTINI D, and NEWMAN P. Get to the point: Learning lidar place recognition and metric localisation using overhead imagery[C]. Proceedings of Robotics: Science and Systems 2021, 2021. doi: 10.15607/RSS.2021.XVII.003.(查阅网上资料,未找到出版地信息,请补充).
    [48] TANG T Y, DE MARTINI D, and NEWMAN P. Point-based metric and topological localisation between lidar and overhead imagery[J]. Autonomous Robots, 2023, 47(5): 595–615. doi: 10.1007/s10514-023-10085-w.
    [49] GE Chongjian, CHEN Junsong, XIE Enze, et al. MetaBEV: Solving sensor failures for 3D detection and map segmentation[C]. Proceedings of 2023 IEEE/CVF International Conference on Computer Vision, Paris, France, 2023: 8687–8697. doi: 10.1109/ICCV51070.2023.00801.
    [50] LIU Liu and LI Hongdong. Lending orientation to neural networks for cross-view geo-localization[C]. Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, USA, 2019: 5617–5626. doi: 10.1109/CVPR.2019.00577.
    [51] MADDERN W, PASCOE G, LINEGAR C, et al. 1 year, 1000 km: The oxford robotcar dataset[J]. The International Journal of Robotics Research, 2017, 36(1): 3–15. doi: 10.1177/0278364916679498.
    [52] MADDERN W, PASCOE G, GADD M, et al. Real-time kinematic ground truth for the oxford robotcar dataset[J]. arXiv: 2002.10152, 2020. doi: 10.48550/arXiv.2002.10152.(查阅网上资料,未能确认文献类型,请确认).
    [53] BARNES D, GADD M, MURCUTT P, et al. The oxford radar robotcar dataset: A radar extension to the oxford RobotCar dataset[C]. Proceedings of 2020 IEEE International Conference on Robotics and Automation, Paris, France, 2020: 6433–6438. doi: 10.1109/ICRA40945.2020.9196884.
    [54] AGARWAL S, VORA A, PANDEY G, et al. Ford multi-AV seasonal dataset[J]. The International Journal of Robotics Research, 2020, 39(12): 1367–1376. doi: 10.1177/0278364920961451.
    [55] YANG Hongji, LU Xiufan, and ZHU Yingying. Cross-view geo-localization with layer-to-layer transformer[C]. Proceedings of the 35th International Conference on Neural Information Processing Systems, 2021: 2222. doi: 10.5555/3540261.3542483. (查阅网上资料,未找到出版地信息,请补充).
    [56] SHI Yujiao, YU Xin, CAMPBELL D, et al. Where am I looking at? Joint location and orientation estimation by cross-view matching[C]. Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2020: 4063–4071. doi: 10.1109/CVPR42600.2020.00412.
    [57] REGMI K and SHAH M. Bridging the domain gap for ground-to-aerial image matching[C]. Proceedings of 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Korea (South), 2019: 470–479. doi: 10.1109/ICCV.2019.00056.
    [58] CHENG Lei, WANG Teng, LI Jiawen, et al. Offset regression enhanced cross-view feature interaction for ground-to-aerial geo-localization[J]. IEEE Transactions on Intelligent Vehicles, 2025, 10(1): 205–216. doi: 10.1109/TIV.2024.3411098.
    [59] ZHANG Xiaohan, LI Xingyu, SULTANI W, et al. GeoDTR+: Toward generic cross-view geolocalization via geometric disentanglement[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46(12): 10419–10433. doi: 10.1109/TPAMI.2024.3443652.
    [60] LI Chaoran, YAN Chao, XIANG Xiaojia, et al. AMPLE: Automatic progressive learning for orientation unknown ground-to-aerial geo-localization[J]. IEEE Transactions on Geoscience and Remote Sensing, 2025, 63: 5800115. doi: 10.1109/TGRS.2024.3517654.
    [61] SHUGAEV M, SEMENOV I, ASHLEY K, et al. ArcGeo: Localizing limited field-of-view images using cross-view matching[C]. Proceedings of 2024 IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, USA, 2024: 208–217. doi: 10.1109/WACV57701.2024.00028.
    [62] LI Chaoran, YAN Chao, XIANG Xiaojia, et al. HADGEO: Image based 3-DoF cross-view geo-localization with hard sample mining[C]. Proceedings of 2024 IEEE International Conference on Acoustics, Speech and Signal Processing, Seoul, Korea, Republic of, 2024: 3520–3524. doi: 10.1109/ICASSP48485.2024.10445839.
    [63] YE Junyan, LV Zhutao, LI Weijia, et al. Cross-view image geo-localization with panorama-BEV Co-retrieval network[C]. Proceedings of the 18th European Conference on Computer Vision, Milan, Italy, 2024: 74–90. doi: 10.1007/978-3-031-72913-3_5.
    [64] ZHU Yingying, YANG Hongji, LU Yuxin, et al. Simple, effective and general: A new backbone for cross-view image geo-localization[J]. arXiv: 2302.01572, 2023. doi: 10.48550/arXiv.2302.01572.(查阅网上资料,未能确认文献类型,请确认).
    [65] PARK J, SUNG C, LEE S, et al. Cross-view geo-localization via effective negative sampling[C]. Proceedings of the 2024 24th International Conference on Control, Automation and Systems, Jeju, Korea, Republic of, 2024: 1078–1083. doi: 10.23919/ICCAS63016.2024.10773330.
    [66] ZHANG Xiaohan, SULTANI W, and WSHAH S. Cross-view image sequence geo-localization[C]. Proceedings of 2023 IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, USA, 2023: 2913–2922. doi: 10.1109/WACV56688.2023.00293.
    [67] SHI Yujiao and LI Hongdong. Beyond cross-view image retrieval: Highly accurate vehicle localization using satellite image[C]. Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, USA, 2022: 16989–16999. doi: 10.1109/CVPR52688.2022.01650.
    [68] LENTSCH T, XIA Zimin, CAESAR H, et al. SliceMatch: Geometry-guided aggregation for cross-view pose estimation[C]. Proceedings of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, Canada, 2023: 17225–17234. doi: 10.1109/CVPR52729.2023.01652.
    [69] XIA Zimin, BOOIJ O, and KOOIJ J F P. Convolutional cross-view pose estimation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46(5): 3813–3831. doi: 10.1109/TPAMI.2023.3346924.
    [70] WANG Xiaolong, XU Runsen, CUI Zuofan, et al. Fine-grained cross-view geo-localization using a correlation-aware homography estimator[C]. Proceedings of the 37th International Conference on Neural Information Processing Systems, New Orleans, USA, 2023: 234. doi: 10.5555/3666122.3666356.
    [71] WANG Qiwei, WU Shaoxun, and SHI Yujiao. BevSplat: Resolving height ambiguity via feature-based Gaussian primitives for weakly-supervised cross-view localization[J]. arXiv: 2502.09080, 2025. doi: 10.48550/arXiv.2502.09080.(查阅网上资料,未能确认文献类型,请确认).
    [72] HU Wenmiao, ZHANG Yichen, LIANG Yuxuan, et al. PetalView: Fine-grained location and orientation extraction of street-view images via cross-view local search[C]. Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, Canada, 2023: 56–66. doi: 10.1145/3581783.3612007.
    [73] TANG T Y, DE MARTINI D, WU Shangzhe, et al. Self-supervised localisation between range sensors and overhead imagery[C]. Proceedings of Robotics: Science and Systems, Corvalis, USA, 2020. doi: 10.15607/RSS.2020.XVI.057.
    [74] HU Di, ZHANG Kai, YUAN Xia, et al. Real-time road intersection detection in sparse point cloud based on augmented viewpoints beam model[J]. Sensors, 2023, 23(21): 8854. doi: 10.3390/s23218854.
  • 加载中
图(7) / 表(5)
计量
  • 文章访问数:  44
  • HTML全文浏览量:  17
  • PDF下载量:  7
  • 被引次数: 0
出版历程
  • 录用日期:  2025-12-17
  • 网络出版日期:  2025-12-25

目录

    /

    返回文章
    返回