广义高斯混合过程 / GGMPs: Generalized Gaussian Mixture Processes
1️⃣ 一句话总结
这篇论文提出了一种名为GGMP的新方法,它基于高斯过程,能够有效预测具有多个峰值、不确定性变化和非高斯特征的复杂数据分布,解决了传统方法只能预测单一、简单分布的局限性。
Conditional density estimation is complicated by multimodality, heteroscedasticity, and strong non-Gaussianity. Gaussian processes (GPs) provide a principled nonparametric framework with calibrated uncertainty, but standard GP regression is limited by its unimodal Gaussian predictive form. We introduce the Generalized Gaussian Mixture Process (GGMP), a GP-based method for multimodal conditional density estimation in settings where each input may be associated with a complex output distribution rather than a single scalar response. GGMP combines local Gaussian mixture fitting, cross-input component alignment and per-component heteroscedastic GP training to produce a closed-form Gaussian mixture predictive density. The method is tractable, compatible with standard GP solvers and scalable methods, and avoids the exponentially large latent-assignment structure of naive multimodal GP formulations. Empirically, GGMPs improve distributional approximation on synthetic and real-world datasets with pronounced non-Gaussianity and multimodality.
广义高斯混合过程 / GGMPs: Generalized Gaussian Mixture Processes
这篇论文提出了一种名为GGMP的新方法,它基于高斯过程,能够有效预测具有多个峰值、不确定性变化和非高斯特征的复杂数据分布,解决了传统方法只能预测单一、简单分布的局限性。
源自 arXiv: 2603.10442