菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-25
📄 Abstract - SEGAR: Selective Enhancement for Generative Augmented Reality

Generative world models offer a compelling foundation for augmented-reality (AR) applications: by predicting future image sequences that incorporate deliberate visual edits, they enable temporally coherent, augmented future frames that can be computed ahead of time and cached, avoiding per-frame rendering from scratch in real time. In this work, we present SEGAR, a preliminary framework that combines a diffusion-based world model with a selective correction stage to support this vision. The world model generates augmented future frames with region-specific edits while preserving others, and the correction stage subsequently aligns safety-critical regions with real-world observations while preserving intended augmentations elsewhere. We demonstrate this pipeline in driving scenarios as a representative setting where semantic region structure is well defined and real-world feedback is readily available. We view this as an early step toward generative world models as practical AR infrastructure, where future frames can be generated, cached, and selectively corrected on demand.

顶级标签: multi-modal computer vision aigc
详细标签: generative world models augmented reality diffusion models selective correction video generation 或 搜索:

SEGAR:面向生成式增强现实的选择性增强框架 / SEGAR: Selective Enhancement for Generative Augmented Reality


1️⃣ 一句话总结

这篇论文提出了一个名为SEGAR的初步框架,它结合了扩散世界模型和选择性校正步骤,能够提前生成并缓存带有特定区域视觉编辑的未来增强现实画面,同时确保安全关键区域与真实世界保持一致,从而为实现高效、可靠的生成式增强现实应用迈出了早期的一步。

源自 arXiv: 2603.24541