菜单

🤖 系统
📄 Abstract - RefVTON: person-to-person Try on with Additional Unpaired Visual Reference

We introduce RefTON, a flux-based person-to-person virtual try-on framework that enhances garment realism through unpaired visual references. Unlike conventional approaches that rely on complex auxiliary inputs such as body parsing and warped mask or require finely designed extract branches to process various input conditions, RefTON streamlines the process by directly generating try-on results from a source image and a target garment, without the need for structural guidance or auxiliary components to handle diverse inputs. Moreover, inspired by human clothing selection behavior, RefTON leverages additional reference images (the target garment worn on different individuals) to provide powerful guidance for refining texture alignment and maintaining the garment details. To enable this capability, we built a dataset containing unpaired reference images for training. Extensive experiments on public benchmarks demonstrate that RefTON achieves competitive or superior performance compared to state-of-the-art methods, while maintaining a simple and efficient person-to-person design.

顶级标签: computer vision aigc multi-modal
详细标签: virtual try-on image generation garment transfer reference guidance diffusion models 或 搜索:

📄 论文总结

RefVTON:基于额外非配对视觉参考的人对人虚拟试穿 / RefVTON: person-to-person Try on with Additional Unpaired Visual Reference


1️⃣ 一句话总结

这篇论文提出了一个名为RefTON的虚拟试穿系统,它通过引入不同人穿着目标服装的参考图片来提升试穿效果的真实感和细节还原,同时简化了传统方法中复杂的输入要求,实现了高效且高质量的人对人服装替换。


📄 打开原文 PDF