菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-31
📄 Abstract - End-to-End Image Compression with Segmentation Guided Dual Coding for Wind Turbines

Transferring large volumes of high-resolution images during wind turbine inspections introduces a bottleneck in assessing and detecting severe defects. Efficient coding must preserve high fidelity in blade regions while aggressively compressing the background. In this work, we propose an end-to-end deep learning framework that jointly performs segmentation and dual-mode (lossy and lossless) compression. The segmentation module accurately identifies the blade region, after which our region-of-interest (ROI) compressor encodes it at superior quality compared to the rest of the image. Unlike conventional ROI schemes that merely allocate more bits to salient areas, our framework integrates: (i) a robust segmentation network (BU-Netv2+P) with a CRF-regularized loss for precise blade localization, (ii) a hyperprior-based autoencoder optimized for lossy compression, and (iii) an extended bits-back coder with hierarchical models for fully lossless blade reconstruction. Furthermore, our ROI framework removes the sequential dependency in bits-back coding by reusing background-coded bits, enabling parallelized and efficient dual-mode compression. To the best of our knowledge, this is the first fully integrated learning-based ROI codec combining segmentation, lossy, and lossless compression, ensuring that subsequent defect detection is not compromised. Experiments on a large-scale wind turbine dataset demonstrate superior compression performance and efficiency, offering a practical solution for automated inspections.

顶级标签: computer vision systems model training
详细标签: image compression segmentation region-of-interest lossless compression wind turbine inspection 或 搜索:

面向风力涡轮机的端到端图像压缩:基于分割引导的双重编码 / End-to-End Image Compression with Segmentation Guided Dual Coding for Wind Turbines


1️⃣ 一句话总结

这篇论文提出了一种用于风力涡轮机巡检图像的新型智能压缩方法,它通过自动识别叶片区域,对关键部分进行高质量压缩,对背景进行高压缩率处理,从而在保证后续缺陷检测精度的同时,大幅提升图像传输和存储效率。

源自 arXiv: 2603.29927