菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-09
📄 Abstract - Fair Feature Importance Scores via Feature Occlusion and Permutation

As machine learning models increasingly impact society, their opaque nature poses challenges to trust and accountability, particularly in fairness contexts. Understanding how individual features influence model outcomes is crucial for building interpretable and equitable models. While feature importance metrics for accuracy are well-established, methods for assessing feature contributions to fairness remain underexplored. We propose two model-agnostic approaches to measure fair feature importance. First, we propose to compare model fairness before and after permuting feature values. This simple intervention-based approach decouples a feature and model predictions to measure its contribution to training. Second, we evaluate the fairness of models trained with and without a given feature. This occlusion-based score enjoys dramatic computational simplification via minipatch learning. Our empirical results reflect the simplicity and effectiveness of our proposed metrics for multiple predictive tasks. Both methods offer simple, scalable, and interpretable solutions to quantify the influence of features on fairness, providing new tools for responsible machine learning development.

顶级标签: model evaluation machine learning theory
详细标签: feature importance fairness interpretability model-agnostic responsible ai 或 搜索:

通过特征遮挡与置换实现公平的特征重要性评分 / Fair Feature Importance Scores via Feature Occlusion and Permutation


1️⃣ 一句话总结

这篇论文提出了两种与模型无关的方法,通过置换或遮挡特征来量化每个特征对机器学习模型公平性的影响,为开发更负责任、可解释的公平模型提供了简单有效的工具。

源自 arXiv: 2602.09196