We consider the problem of the stability of saliency-based explanations of
Neural Network predictions under adversarial attacks in a classification task.
Saliency interpretations of deterministic Neural Networks are remarkably
brittle even when the attacks fail, i.e. for attacks that do not change the
classification label. We empirically show that interpretations provided by
Bayesian Neural Networks are considerably more stable under adversarial
perturbations of the inputs and even under direct attacks to the explanations.
By leveraging recent results, we also provide a theoretical explanation of this
result in terms of the geometry of the data manifold. Additionally, we discuss
the stability of the interpretations of high level representations of the
inputs in the internal layers of a Network. Our results demonstrate that
Bayesian methods, in addition to being more robust to adversarial attacks, have
the potential to provide more stable and interpretable assessments of Neural
Network predictions.