MapSR:通过视觉基础模型实现提示驱动的地表覆盖图超分辨率 / MapSR: Prompt-Driven Land Cover Map Super-Resolution via Vision Foundation Models
1️⃣ 一句话总结
这篇论文提出了一个名为MapSR的新方法,它利用预训练好的视觉基础模型,仅需一次低分辨率标签就能快速生成高分辨率的地表覆盖图,无需大量标注数据和长时间训练,大大降低了计算成本。
High-resolution (HR) land-cover mapping is often constrained by the high cost of dense HR annotations. We revisit this problem from the perspective of map super-resolution, which enhances coarse low-resolution (LR) land-cover products into HR maps at the resolution of the input imagery. Existing weakly supervised methods can leverage LR labels, but they typically use them to retrain dense predictors with substantial computational cost. We propose MapSR, a prompt-driven framework that decouples supervision from model training. MapSR uses LR labels once to extract class prompts from frozen vision foundation model features through a lightweight linear probe, after which HR mapping proceeds via training-free metric inference and graph-based prediction refinement. Specifically, class prompts are estimated by aggregating high-confidence HR features identified by the linear probe, and HR predictions are obtained by cosine-similarity matching followed by graph-based propagation for spatial refinement. Experiments on the Chesapeake Bay dataset show that MapSR achieves 59.64% mIoU without any HR labels, remaining competitive with the strongest weakly supervised baseline and surpassing a fully supervised baseline. Notably, MapSR reduces trainable parameters by four orders of magnitude and shortens training time from hours to minutes, enabling scalable HR mapping under limited annotation and compute budgets. The code is available at this https URL.
MapSR:通过视觉基础模型实现提示驱动的地表覆盖图超分辨率 / MapSR: Prompt-Driven Land Cover Map Super-Resolution via Vision Foundation Models
这篇论文提出了一个名为MapSR的新方法,它利用预训练好的视觉基础模型,仅需一次低分辨率标签就能快速生成高分辨率的地表覆盖图,无需大量标注数据和长时间训练,大大降低了计算成本。
源自 arXiv: 2604.14582