These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Black-box machine learning models are being used in more and more high-stakes
domains, which creates a growing need for Explainable AI (XAI). Unfortunately,
the use of XAI in machine learning introduces new privacy risks, which
currently remain largely unnoticed. We introduce the explanation linkage
attack, which can occur when deploying instance-based strategies to find
counterfactual explanations. To counter such an attack, we propose k-anonymous
counterfactual explanations and introduce pureness as a new metric to evaluate
the validity of these k-anonymous counterfactual explanations. Our results show
that making the explanations, rather than the whole dataset, k- anonymous, is
beneficial for the quality of the explanations.