基于混合2D-3D CNN框架的低采样3D锥束CT图像伪影减少方法 / Artifact Reduction in Undersampled 3D Cone-Beam CTs using a Hybrid 2D-3D CNN Framework
1️⃣ 一句话总结
这篇论文提出了一种结合2D和3D卷积神经网络优势的混合深度学习框架,能够高效地减少低采样3D CT图像中的伪影,在显著提升图像质量和切片间一致性的同时,保持了较低的计算成本。
Undersampled CT volumes minimize acquisition time and radiation exposure but introduce artifacts degrading image quality and diagnostic utility. Reducing these artifacts is critical for high-quality imaging. We propose a computationally efficient hybrid deep-learning framework that combines the strengths of 2D and 3D models. First, a 2D U-Net operates on individual slices of undersampled CT volumes to extract feature maps. These slice-wise feature maps are then stacked across the volume and used as input to a 3D decoder, which utilizes contextual information across slices to predict an artifact-free 3D CT volume. The proposed two-stage approach balances the computational efficiency of 2D processing with the volumetric consistency provided by 3D modeling. The results show substantial improvements in inter-slice consistency in coronal and sagittal direction with low computational overhead. This hybrid framework presents a robust and efficient solution for high-quality 3D CT image post-processing. The code of this project can be found on github: this https URL.
基于混合2D-3D CNN框架的低采样3D锥束CT图像伪影减少方法 / Artifact Reduction in Undersampled 3D Cone-Beam CTs using a Hybrid 2D-3D CNN Framework
这篇论文提出了一种结合2D和3D卷积神经网络优势的混合深度学习框架,能够高效地减少低采样3D CT图像中的伪影,在显著提升图像质量和切片间一致性的同时,保持了较低的计算成本。
源自 arXiv: 2602.08727