基于优化的机器学习分类器逻辑解释泛化方法 / Generalizing Logic-based Explanations for Machine Learning Classifiers via Optimization
1️⃣ 一句话总结
这篇论文提出了两种新的优化方法,通过一次性或分步调整特征边界,在保证解释正确性的前提下,显著提高了逻辑解释的覆盖范围,从而让机器学习模型的决策依据更可靠、更通用。
Machine learning models support decision-making, yet the reasons behind their predictions are opaque. Clear and reliable explanations help users make informed decisions and avoid blindly trusting model outputs. However, many existing explanation methods fail to guarantee correctness. Logic-based approaches ensure correctness but often offer overly constrained explanations, limiting coverage. Recent work addresses this by incrementally expanding explanations while maintaining correctness. This process is performed separately for each feature, adjusting both its upper and lower bounds. However, this approach faces a trade-off: smaller increments incur high computational costs, whereas larger ones may lead to explanations covering fewer instances. To overcome this, we propose two novel methods. Onestep builds upon this prior work, generating explanations in a single step for each feature and each bound, eliminating the overhead of an iterative process. \textit{Twostep} takes a gradual approach, improving coverage. Experimental results show that Twostep significantly increases explanation coverage (by up to 72.60\% on average across datasets) compared to Onestep and, consequently, to prior work.
基于优化的机器学习分类器逻辑解释泛化方法 / Generalizing Logic-based Explanations for Machine Learning Classifiers via Optimization
这篇论文提出了两种新的优化方法,通过一次性或分步调整特征边界,在保证解释正确性的前提下,显著提高了逻辑解释的覆盖范围,从而让机器学习模型的决策依据更可靠、更通用。
源自 arXiv: 2603.01870