菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-05
📄 Abstract - Measuring the Fragility of Trust: Devising Credibility Index via Explanation Stability (CIES) for Business Decision Support Systems

Explainable Artificial Intelligence (XAI) methods (SHAP, LIME) are increasingly adopted to interpret models in high-stakes businesses. However, the credibility of these explanations, their stability under realistic data perturbations, remains unquantified. This paper introduces the Credibility Index via Explanation Stability (CIES), a mathematically grounded metric that measures how robust a model's explanations are when subject to realistic business noise. CIES captures whether the reasons behind a prediction remain consistent, not just the prediction itself. The metric employs a rank-weighted distance function that penalizes instability in the most important features disproportionately, reflecting business semantics where changes in top decision drivers are more consequential than changes in marginal features. We evaluate CIES across three datasets (customer churn, credit risk, employee attrition), four tree-based classification models and two data balancing conditions. Results demonstrate that model complexity impacts explanation credibility, class imbalance treatment via SMOTE affects not only predictive performance but also explanation stability, and CIES provides statistically superior discriminative power compared to a uniform baseline metric (p < 0.01 in all 24 configurations). A sensitivity analysis across four noise levels confirms the robustness of the metric itself. These findings offer business practitioners a deployable &#34;credibility warning system&#34; for AI-driven decision support.

顶级标签: model evaluation machine learning systems
详细标签: explainable ai trust measurement robustness business ai stability metric 或 搜索:

衡量信任的脆弱性:通过解释稳定性构建可信度指数以支持商业决策系统 / Measuring the Fragility of Trust: Devising Credibility Index via Explanation Stability (CIES) for Business Decision Support Systems


1️⃣ 一句话总结

这篇论文提出了一种名为CIES的新指标,用于量化人工智能模型解释的可信度,它通过测试解释在数据扰动下的稳定性,帮助商业用户判断AI决策建议是否可靠。

源自 arXiv: 2603.05024