高级搜索

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

面向稀疏辐射观测的无监督三维医学图像分割方法

俞晓帆 邹兰兰 顾文琦 蔡君 康彬 丁康

俞晓帆, 邹兰兰, 顾文琦, 蔡君, 康彬, 丁康. 面向稀疏辐射观测的无监督三维医学图像分割方法[J]. 电子与信息学报. doi: 10.11999/JEIT250841
引用本文: 俞晓帆, 邹兰兰, 顾文琦, 蔡君, 康彬, 丁康. 面向稀疏辐射观测的无监督三维医学图像分割方法[J]. 电子与信息学报. doi: 10.11999/JEIT250841
YU Xiaofan, ZOU Lanlan, GU Wenqi, CAI Jun, KANG Bin, DING Kang. Unsupervised 3D Medical Image Segmentation With Sparse Radiation Measurement[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT250841
Citation: YU Xiaofan, ZOU Lanlan, GU Wenqi, CAI Jun, KANG Bin, DING Kang. Unsupervised 3D Medical Image Segmentation With Sparse Radiation Measurement[J]. Journal of Electronics & Information Technology. doi: 10.11999/JEIT250841

面向稀疏辐射观测的无监督三维医学图像分割方法

doi: 10.11999/JEIT250841 cstr: 32379.14.JEIT250841
基金项目: 国家自然科学基金(62171232),江苏省中医肛肠疾病临床医学创新中心重点项目(GCCXZX-2021),南京市卫生科技发展专项资金重点项目(ZKX24046),国家中医药管理局监测统计中心2025 年度中医药监测统计研究课题(2025JCTJE2),国家中医药管理局规划财务司项目(GZY-GCS-2025-006)
详细信息
    作者简介:

    俞晓帆:男,教授,研究方向为智能信号处理

    邹兰兰:女,硕士生,研究方向为医学图像处理

    顾文琦:女,硕士生,研究方向为计算机视觉

    蔡君:女,讲师,研究方向为医学信号处理

    康彬:男,副教授,研究方向为计算机视觉

    丁康:男,教授,研究方向为肛肠病临床研究

    通讯作者:

    丁康 fsyy00237@njucm.edu.cn

Unsupervised 3D Medical Image Segmentation With Sparse Radiation Measurement

Funds: The National Natural Science Foundation of China (62171232), The Key Projects of the Jiangsu Provincial Clinical Medicine Innovation Center for Anorectal Diseases (GC-CXZX-2021), The Key Projects of the Nanjing Municipal Special Fund for Health Science and Technology Development (ZKX24046), The TCM Monitoring and Statistics from the National Administration of Traditional Chinese Medicine Statistics Center (2025JCTJE2), National Administration of Traditional Chinese Medicine Planning and Finance Department Project (GZY-GCS-2025-006)
  • 摘要: 神经衰减场是一种具有前景的三维医学图像重建方法,此方法利用稀疏辐射测量实现与完整观察相接近的重构精度。该文提出一种无监督三维医学影像分割方法,将无监督分割与神经衰减场集成为一个端到端的网络架构。具体而言,所提网络架构包括两个阶段:稀疏测量重建和交互式三维图像分割。两个阶段可通过联合学习自适应实现互惠优化。为解决类似肛肠等复杂病灶中边界模糊和区域过度扩展的难题,所提三维分割网络的交互式三维分割阶段设计了密度引导模块,有效利用衰减系数的先验知识,调节密度感知的注意力机制,提升三维分割泛化性能。通过与南京市中医院合作构建的结直肠癌数据集以及两个公开数据集上的大量实验证明所提方法的优越性,例如与基于全辐射观测的SAM-MED3D算法相比,所提网络仅使用14%稀疏观测值,在3个数据集的平均 Dice 系数提升 2.0%。
  • 图  1  所提无监督三维分割方法与传统监督分割方法对比示例

    图  2  NA-SAM3D 框架总体结构。该框架包含两个主要阶段:(1)基于稀疏视图的三维重建;(2)交互式三维分割。其中密度引导模块为分割阶段的关键组成部分,用于增强边界区域的识别能力。

    图  3  不同分割方法在自建结直肠癌数据集上的二维切片结果对比

    图  4  不同分割方法在公共数据集上的二维切片结果对比

    图  5  不同分割方法在三维可视化结果上的对比

    图  6  在引入与未引入密度引导模块条件下,不同数据集的Dice与mIoU性能对比

    图  7  不同交互点数量下模型平均 Dice 值和方差的变化趋势

    图  8  不同稀疏视角数量条件下的平均 Dice 值和方差的变化趋势

    图  9  不同采样点数量条件下的平均 Dice 值和方差的变化趋势

    表  1  不同数据集的定量结果对比,(表中加粗值表示该列指标的最优结果,HD95 与 ASD 的单位均为 mm)

    方法LCTSCLiTS自建数据集
    DICEmIoUHD95ASDDICEmIoUHD95ASDDICEmIoUHD95ASD
    Swinunetr-v2[5]0.78120.628518.543.660.88650.803310.363.270.75980.672117.283.71
    UNETR++[6]0.81710.700311.203.040.91260.831210.083.230.78350.68949.702.75
    SAM-MED2D[18]0.58050.409024.895.130.86230.757712.943.860.71960.636126.885.40
    SA3D[21]0.69310.538923.484.910.80270.712326.815.920.72350.610825.145.73
    SAM-MED3D[22]0.76230.625721.214.690.85110.774117.004.220.74480.642618.254.05
    本文0.79460.629615.923.370.86290.792412.793.310.76100.673313.613.05
    下载: 导出CSV

    表  2  不同模型的均值与方差比较(表中加粗值表示该列指标的最优结果)

    方法平均值方差
    DICEmIoUHD95ASDDICEmIoUHD95ASD
    Swinunetr-v2[5]0.80920.701315.393.550.00310.005512.930.04
    UNETR++[6]0.83770.740310.332.940.00300.00420.410.02
    SAM-MED2D[18]0.72080.600921.574.800.01320.020937.900.45
    SA3D[21]0.73980.620725.145.520.00210.00511.840.19
    SAM-MED3D[22]0.78610.680818.824.320.00220.00443.110.07
    本文0.80610.698414.113.240.00180.00471.760.02
    下载: 导出CSV

    表  3  基于自建数据集微调下的交叉域评估结果,(表中加粗值表示该列指标的最优结果)

    方法LCTSCLiTS
    DICEmIoUHD95ASDDICEmIoUHD95ASD
    SAM-MED2D[18]0.46450.367227.217.960.67530.578228.907.22
    SAM-MED3D[22]0.70180.562223.835.760.69210.636822.826.38
    本文0.74420.589418.064.550.72310.654916.124.71
    下载: 导出CSV

    表  4  不同重建方法实验结果对比

    三维重建
    方法
    正常投影
    DICE (avg)
    正常投影
    mIoU (avg)
    加入噪声投影
    DICE (avg)
    加入噪声投影
    mIoU (avg)
    NAF[12] 0.8061 0.6984 0.7935 0.6811
    FDK[25] 0.7351 0.5966 0.7083 0.5644
    SART[26] 0.7168 0.5863 0.6791 0.5580
    NeRF[11] 0.7550 0.6144 0.7335 0.5916
    下载: 导出CSV
  • [1] EMRE T, CHAKRAVARTY A, RIVAIL A, et al. 3DTINC: Time-equivariant non-contrastive learning for predicting disease progression from longitudinal OCTs[J]. IEEE Transactions on Medical Imaging, 2024, 43(9): 3200–3210. doi: 10.1109/TMI.2024.3391215.
    [2] ZHU Pengfei, WANG Tingmin, YANG Fan, et al. A transformer-based multi-scale deep learning model for lung cancer surgery optimization[J]. IEEE Access, 2025, 13: 70044–70054. doi: 10.1109/ACCESS.2025.3561948.
    [3] WU Junde, ZHANG Yu, FANG Huihui, et al. Calibrate the inter-observer segmentation uncertainty via diagnosis-first principle[J]. IEEE Transactions on Medical Imaging, 2024, 43(9): 3331–3342. doi: 10.1109/TMI.2024.3394045.
    [4] MA Yuxi, WANG Jiacheng, YANG Jing, et al. Model-heterogeneous semi-supervised federated learning for medical image segmentation[J]. IEEE Transactions on Medical Imaging, 2024, 43(5): 1804–1815. doi: 10.1109/TMI.2023.3348982.
    [5] HE Yufan, NATH V, YANG Dong, et al. SwinUNETR-V2: Stronger swin transformers with stagewise convolutions for 3D medical image segmentation[C]. The 26th International Conference on Medical Image Computing and Computer Assisted Intervention, Vancouver, Canada, 2023: 416–426. doi: 10.1007/978-3-031-43901-8_40.
    [6] SHAKER A, MAAZ M, RASHEED H, et al. UNETR++: Delving into efficient and accurate 3D medical image segmentation[J]. IEEE Transactions on Medical Imaging, 2024, 43(9): 3377–3390. doi: 10.1109/TMI.2024.3398728.
    [7] PERERA S, NAVARD P, YILMAZ A, et al. SegFormer3D: An efficient transformer for 3D medical image segmentation[C]. IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, USA, 2024: 4981–4988. doi: 10.1109/CVPRW63382.2024.00503.
    [8] TOWLE B, CHEN Xin, and ZHOU Ke. SimSAM: Zero-shot medical image segmentation via simulated interaction[C]. The 2024 IEEE International Symposium on Biomedical Imaging, Athens, Greece, 2024: 1–5. doi: 10.1109/ISBI56570.2024.10635227.
    [9] BUI N T, HOANG D H, TRAN M T, et al. SAM3D: Segment anything model in volumetric medical images[C]. 2024 IEEE International Symposium on Biomedical Imaging, Athens, Greece, 2024: 1–4. doi: 10.1109/ISBI56570.2024.10635844.
    [10] YAMAGISHI Y, HANAOKA S, KIKUCHI T, et al. Using segment anything model 2 for zero-shot 3D segmentation of abdominal organs in computed tomography scans to adapt video tracking capabilities for 3D medical imaging: Algorithm development and validation[J]. JMIR AI, 2025, 4(1): e72109. doi: 10.2196/72109.
    [11] MILDENHALL B, SRINIVASAN P P, TANCIK M, et al. NeRF: Representing scenes as neural radiance fields for view synthesis[J]. Communications of the ACM, 2021, 65(1): 99–106. doi: 10.1145/3503250.
    [12] ZHA Ruyi, ZHANG Yanhao, LI Hongdong, et al. NAF: Neural attenuation fields for sparse-view CBCT reconstruction[C]. The 25th International Conference on Medical Image Computing and Computer Assisted Intervention, Singapore, 2022: 442–452. doi: 10.1007/978-3-031-16446-0_42.
    [13] TAN Pinhuang, GENG Mengxiao, LU Jingya, et al. MSDiff: Multi-scale diffusion model for ultra-sparse view CT reconstruction[J]. arXiv preprint arXiv: 2405.05814, 2024. doi: 10.48550/arXiv.2405.05814.
    [14] SONG Bowen, HU J, LUO Zhaoxu, et al. DiffusionBlend: Learning 3D image prior through position-aware diffusion score blending for 3D computed tomography reconstruction[C]. The 38th International Conference on Neural Information Processing Systems, Vancouver, Canada, 2024: 2844.
    [15] ZHAO Xuzhi, DU Yi, YUE Haizhen, et al. Deep learning-based projection synthesis for low-dose cone-beam computed tomography imaging in image-guided radiotherapy[J]. Quantitative Imaging in Medicine and Surgery, 2024, 14(1): 23150–23250. doi: 10.21037/qims-23-759.
    [16] YANG Liutao, HUANG Jiahao, YANG Guang, et al. CT-SDM: A sampling diffusion model for sparse-view CT reconstruction across various sampling rates[J]. IEEE Transactions on Medical Imaging, 2025, 44(6): 2581–2593. doi: 10.1109/TMI.2025.3541491.
    [17] KANG E, CHANG W, YOO J, et al. Deep convolutional framelet denosing for low-dose CT via wavelet residual network[J]. IEEE Transactions on Medical Imaging, 2018, 37(6): 1358–1369. doi: 10.1109/TMI.2018.2823756.
    [18] CHENG Junlong, YE Jin, DENG Zhongying, et al. SAM-Med2D[J]. arXiv preprint arXiv: 2308.16184, 2023. doi: 10.48550/arXiv.2308.16184.
    [19] LEI Wenhui, XU Wei, LI Kang, et al. MedLSAM: Localize and segment anything model for 3D CT images[J]. Medical Image Analysis, 2025, 99: 103370. doi: 10.1016/j.media.2024.103370.
    [20] GONG Shizhan, ZHONG Yuan, MA Wenao, et al. 3DSAM-Adapter: Holistic adaptation of SAM from 2D to 3D for promptable tumor segmentation[J]. Medical Image Analysis, 2024, 98: 103324. doi: 10.1016/j.media.2024.103324.
    [21] CEN Jiazhong, FANG Jiemin, ZHOU Zanwei, et al. Segment anything in 3D with radiance fields[J]. International Journal of Computer Vision, 2025, 133(8): 5138–5160. doi: 10.1007/s11263-025-02421-7.
    [22] WANG Haoyu, GUO Sizheng, YE Jin, et al. SAM-Med3D: Towards general-purpose segmentation models for volumetric medical images[C]. The European Conference on Computer Vision, Milan, Italy, 2024: 51–67. doi: 10.1007/978-3-031-91721-9_4.
    [23] YANG J, SHARP G, VEERARAGHAVAN H, et al. Data from Lung CT Segmentation Challenge (LCTSC) (Version 3)[M]. The Cancer Imaging Archive, 2017.
    [24] BILIC P, CHRIST P, LI H B, et al. The Liver Tumor Segmentation benchmark (LiTS)[J]. Medical Image Analysis, 2023, 84: 102680. doi: 10.1016/j.media.2022.102680.
    [25] FELDKAMP L A, DAVIS L C, and KRESS J W. Practical cone-beam algorithm[J]. Journal of the Optical Society of America A, 1984, 1(6): 612–619. doi: 10.1364/JOSAA.1.000612.
    [26] ANDERSEN A H and KAK A C. Simultaneous Algebraic Reconstruction Technique (SART): A superior implementation of the ART algorithm[J]. Ultrasonic Imaging, 1984, 6(1): 81–94. doi: 10.1177/016173468400600107.
    [27] CAI Yuanhao, WANG Jiahao, YUILLE A, et al. Structure-aware sparse-view X-ray 3D reconstruction[C]. IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2024: 11174–11183. doi: 10.1109/CVPR52733.2024.01062.
  • 加载中
图(9) / 表(4)
计量
  • 文章访问数:  112
  • HTML全文浏览量:  62
  • PDF下载量:  17
  • 被引次数: 0
出版历程
  • 收稿日期:  2025-09-01
  • 修回日期:  2025-11-05
  • 录用日期:  2025-11-05
  • 网络出版日期:  2025-11-15

目录

    /

    返回文章
    返回