菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-14
📄 Abstract - MAST: Mask-Guided Attention Mass Allocation for Training-Free Multi-Style Transfer

Style transfer aims to render a content image with the visual characteristics of a reference style while preserving its underlying semantic layout and structural geometry. While recent diffusion-based models demonstrate strong stylization capabilities by leveraging powerful generative priors and controllable internal representations, they typically assume a single global style. Extending them to multi-style scenarios often leads to boundary artifacts, unstable stylization, and structural inconsistency due to interference between multiple style representations. To overcome these limitations, we propose MAST (Mask-Guided Attention Mass Allocation for Training-Free Multi-Style Transfer), a novel training-free framework that explicitly controls content-style interactions within the diffusion attention mechanism. To achieve artifact-free and structure-preserving stylization, MAST integrates four connected modules. First, Layout-preserving Query Anchoring prevents global layout collapse by firmly anchoring the semantic structure using content queries. Second, Logit-level Attention Mass Allocation deterministically distributes attention probability mass across spatial regions, seamlessly fusing multiple styles without boundary artifacts. Third, Sharpness-aware Temperature Scaling restores the attention sharpness degraded by multi-style expansion. Finally, Discrepancy-aware Detail Injection adaptively compensates for localized high-frequency detail losses by measuring structural discrepancies. Extensive experiments demonstrate that MAST effectively mitigates boundary artifacts and maintains structural consistency, preserving texture fidelity and spatial coherence even as the number of applied styles increases.

顶级标签: computer vision model training multi-modal
详细标签: style transfer diffusion models attention mechanism training-free image generation 或 搜索:

MAST:基于掩码引导注意力质量分配的免训练多风格迁移方法 / MAST: Mask-Guided Attention Mass Allocation for Training-Free Multi-Style Transfer


1️⃣ 一句话总结

这篇论文提出了一种名为MAST的免训练新方法,它通过精确控制扩散模型中的注意力分配,成功解决了多风格图像合成中常见的边界瑕疵和结构失真问题,实现了高质量、无干扰的多风格融合。

源自 arXiv: 2604.12281