菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-25
📄 Abstract - Omni-Weather: Unified Multimodal Foundation Model for Weather Generation and Understanding

Weather modeling requires both accurate prediction and mechanistic interpretation, yet existing methods treat these goals in isolation, separating generation from understanding. To address this gap, we present Omni-Weather, the first multimodal foundation model that unifies weather generation and understanding within a single architecture. Omni-Weather integrates a radar encoder for weather generation tasks, followed by unified processing using a shared self-attention mechanism. Moreover, we construct a Chain-of-Thought dataset for causal reasoning in weather generation, enabling interpretable outputs and improved perceptual quality. Extensive experiments show Omni-Weather achieves state-of-the-art performance in both weather generation and understanding. Our findings further indicate that generative and understanding tasks in the weather domain can mutually enhance each other. Omni-Weather also demonstrates the feasibility and value of unifying weather generation and understanding.

顶级标签: multi-modal model training aigc
详细标签: weather modeling multimodal foundation model causal reasoning chain-of-thought radar encoder 或 搜索:

全能天气:用于天气生成与理解的统一多模态基础模型 / Omni-Weather: Unified Multimodal Foundation Model for Weather Generation and Understanding


1️⃣ 一句话总结

这篇论文提出了首个名为‘Omni-Weather’的统一多模态基础模型,它在一个架构内同时处理天气的生成与理解任务,通过共享机制和因果推理数据集,不仅提升了性能,还证明了这两类任务可以相互促进。

源自 arXiv: 2512.21643