菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-15
📄 Abstract - VibeFlow: Versatile Video Chroma-Lux Editing through Self-Supervised Learning

Video chroma-lux editing, which aims to modify illumination and color while preserving structural and temporal fidelity, remains a significant challenge. Existing methods typically rely on expensive supervised training with synthetic paired data. This paper proposes VibeFlow, a novel self-supervised framework that unleashes the intrinsic physical understanding of pre-trained video generation models. Instead of learning color and light transitions from scratch, we introduce a disentangled data perturbation pipeline that enforces the model to adaptively recombine structure from source videos and color-illumination cues from reference images, enabling robust disentanglement in a self-supervised manner. Furthermore, to rectify discretization errors inherent in flow-based models, we introduce Residual Velocity Fields alongside a Structural Distortion Consistency Regularization, ensuring rigorous structural preservation and temporal coherence. Our framework eliminates the need for costly training resources and generalizes in a zero-shot manner to diverse applications, including video relighting, recoloring, low-light enhancement, day-night translation, and object-specific color editing. Extensive experiments demonstrate that VibeFlow achieves impressive visual quality with significantly reduced computational overhead. Our project is publicly available at this https URL.

顶级标签: computer vision video model training
详细标签: video editing self-supervised learning color transfer temporal coherence zero-shot generalization 或 搜索:

VibeFlow:通过自监督学习实现多功能视频色彩-光照编辑 / VibeFlow: Versatile Video Chroma-Lux Editing through Self-Supervised Learning


1️⃣ 一句话总结

这篇论文提出了一种名为VibeFlow的自监督学习框架,它能够巧妙地利用现有视频生成模型的知识,无需大量人工标注数据,就能高效、高质量地完成视频的重新打光、重新着色、低光增强等多种编辑任务,同时保持视频结构的稳定和播放的流畅。

源自 arXiv: 2604.13425