菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-10
📄 Abstract - OmniPSD: Layered PSD Generation with Diffusion Transformer

Recent advances in diffusion models have greatly improved image generation and editing, yet generating or reconstructing layered PSD files with transparent alpha channels remains highly challenging. We propose OmniPSD, a unified diffusion framework built upon the Flux ecosystem that enables both text-to-PSD generation and image-to-PSD decomposition through in-context learning. For text-to-PSD generation, OmniPSD arranges multiple target layers spatially into a single canvas and learns their compositional relationships through spatial attention, producing semantically coherent and hierarchically structured layers. For image-to-PSD decomposition, it performs iterative in-context editing, progressively extracting and erasing textual and foreground components to reconstruct editable PSD layers from a single flattened image. An RGBA-VAE is employed as an auxiliary representation module to preserve transparency without affecting structure learning. Extensive experiments on our new RGBA-layered dataset demonstrate that OmniPSD achieves high-fidelity generation, structural consistency, and transparency awareness, offering a new paradigm for layered design generation and decomposition with diffusion transformers.

顶级标签: computer vision multi-modal model training
详细标签: diffusion transformer psd generation image decomposition rgba-vae layered design 或 搜索:

OmniPSD:基于扩散变换器的分层PSD生成 / OmniPSD: Layered PSD Generation with Diffusion Transformer


1️⃣ 一句话总结

这篇论文提出了一个名为OmniPSD的统一框架,它利用扩散变换器技术,既能根据文字描述生成结构清晰、带透明通道的分层PSD设计文件,也能将一张普通图片分解成可编辑的PSD图层,为数字设计提供了新的自动化工具。


源自 arXiv: 2512.09247