弥合评估鸿沟:多目标搜索的标准化基准 / Bridging the Evaluation Gap: Standardized Benchmarks for Multi-Objective Search
1️⃣ 一句话总结
这篇论文针对多目标搜索领域缺乏统一评估标准的问题,创建了一个包含多种真实与合成场景的标准化测试套件,旨在让不同算法的性能对比更可靠、可复现且全面。
Empirical evaluation in multi-objective search (MOS) has historically suffered from fragmentation, relying on heterogeneous problem instances with incompatible objective definitions that make cross-study comparisons difficult. This standardization gap is further exacerbated by the realization that DIMACS road networks, a historical default benchmark for the field, exhibit highly correlated objectives that fail to capture diverse Pareto-front structures. To address this, we introduce the first comprehensive, standardized benchmark suite for exact and approximate MOS. Our suite spans four structurally diverse domains: real-world road networks, structured synthetic graphs, game-based grid environments, and high-dimensional robotic motion-planning roadmaps. By providing fixed graph instances, standardized start-goal queries, and both exact and approximate reference Pareto-optimal solution sets, this suite captures a full spectrum of objective interactions: from strongly correlated to strictly independent. Ultimately, this benchmark provides a common foundation to ensure future MOS evaluations are robust, reproducible, and structurally comprehensive.
弥合评估鸿沟:多目标搜索的标准化基准 / Bridging the Evaluation Gap: Standardized Benchmarks for Multi-Objective Search
这篇论文针对多目标搜索领域缺乏统一评估标准的问题,创建了一个包含多种真实与合成场景的标准化测试套件,旨在让不同算法的性能对比更可靠、可复现且全面。
源自 arXiv: 2603.24084