全属性:面向视觉概念个性化的开放词汇属性编码器 / Omni-Attribute: Open-vocabulary Attribute Encoder for Visual Concept Personalization
1️⃣ 一句话总结
这篇论文提出了一种名为‘全属性’的新方法,它能够像‘精准拆解’图片一样,单独学习和控制图片中的特定属性(如人物身份、表情、光线或风格),从而在生成新图片时,只改变用户想要的某个特征而不影响其他部分,解决了现有技术中多种特征混杂导致效果不佳的问题。
Visual concept personalization aims to transfer only specific image attributes, such as identity, expression, lighting, and style, into unseen contexts. However, existing methods rely on holistic embeddings from general-purpose image encoders, which entangle multiple visual factors and make it difficult to isolate a single attribute. This often leads to information leakage and incoherent synthesis. To address this limitation, we introduce Omni-Attribute, the first open-vocabulary image attribute encoder designed to learn high-fidelity, attribute-specific representations. Our approach jointly designs the data and model: (i) we curate semantically linked image pairs annotated with positive and negative attributes to explicitly teach the encoder what to preserve or suppress; and (ii) we adopt a dual-objective training paradigm that balances generative fidelity with contrastive disentanglement. The resulting embeddings prove effective for open-vocabulary attribute retrieval, personalization, and compositional generation, achieving state-of-the-art performance across multiple benchmarks.
全属性:面向视觉概念个性化的开放词汇属性编码器 / Omni-Attribute: Open-vocabulary Attribute Encoder for Visual Concept Personalization
这篇论文提出了一种名为‘全属性’的新方法,它能够像‘精准拆解’图片一样,单独学习和控制图片中的特定属性(如人物身份、表情、光线或风格),从而在生成新图片时,只改变用户想要的某个特征而不影响其他部分,解决了现有技术中多种特征混杂导致效果不佳的问题。
源自 arXiv: 2512.10955