These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
In safety-critical applications such as medical imaging and autonomous
driving, where decisions have profound implications for patient health and road
safety, it is imperative to maintain both high adversarial robustness to
protect against potential adversarial attacks and reliable uncertainty
quantification in decision-making. With extensive research focused on enhancing
adversarial robustness through various forms of adversarial training (AT), a
notable knowledge gap remains concerning the uncertainty inherent in
adversarially trained models. To address this gap, this study investigates the
uncertainty of deep learning models by examining the performance of conformal
prediction (CP) in the context of standard adversarial attacks within the
adversarial defense community. It is first unveiled that existing CP methods do
not produce informative prediction sets under the commonly used
$l_{\infty}$-norm bounded attack if the model is not adversarially trained,
which underpins the importance of adversarial training for CP. Our paper next
demonstrates that the prediction set size (PSS) of CP using adversarially
trained models with AT variants is often worse than using standard AT,
inspiring us to research into CP-efficient AT for improved PSS. We propose to
optimize a Beta-weighting loss with an entropy minimization regularizer during
AT to improve CP-efficiency, where the Beta-weighting loss is shown to be an
upper bound of PSS at the population level by our theoretical analysis.
Moreover, our empirical study on four image classification datasets across
three popular AT baselines validates the effectiveness of the proposed
Uncertainty-Reducing AT (AT-UR).
External Datasets
CIFAR10
CIFAR100
Caltech256
CUB200
References
International Conference on Learning Representations
Uncertainty sets for image classifiers using conformal prediction
IEEE transactions on neural networks and learning systems
Accelerating monte carlo bayesian prediction via approximating predictive uncertainty over the simplex
Cui, Y., Yao, W., Li, Q., Chan, A. B., Xue, C. J.
Published: 2020
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Bayesian nested neural networks for uncertainty calibration and adaptive compression
Cui, Y., Liu, Z., Li, Q., Chan, A. B., Xue, C. J.
Published: 2021
The Eleventh International Conference on Learning Representations
Bayes-MIL: A new probabilistic perspective on attention-based multiple instance learning for whole slide images
CUI, Y., Liu, Z., Liu, X., Liu, X., Wang, C., Kuo, T.-W., Xue, C. J., Chan, A. B.
Published: 2023
IEEE Transactions on Pattern Analysis and Machine Intelligence
Variational nested dropout
Cui, Y., Mao, Y., Liu, Z., Li, Q., Chan, A. B., Liu, X., Kuo, T.-W., Xue, C. J.
Published: 2023
Advances in Neural Information Processing Systems
Training uncertainty-aware classifiers with conformalized deep learning
Einbinder, B.-S., Romano, Y., Sesia, M., Zhou, Y.
Published: 2022
arxiv
Cited by 6
International Conference on Machine Learning (ICML)
Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning
Yarin Gal, Zoubin Ghahramani
Published: 6.6.2015
Deep learning tools have gained tremendous attention in applied machine
learning. However such tools for regression and classification do not capture
model uncertainty. In comparison, Bayesian models offer a mathematically
grounded framework to reason about model uncertainty, but usually come with a
prohibitive computational cost. In this paper we develop a new theoretical
framework casting dropout training in deep neural networks (NNs) as approximate
Bayesian inference in deep Gaussian processes. A direct result of this theory
gives us tools to model uncertainty with dropout NNs -- extracting information
from existing models that has been thrown away so far. This mitigates the
problem of representing uncertainty in deep learning without sacrificing either
computational complexity or test accuracy. We perform an extensive study of the
properties of dropout's uncertainty. Various network architectures and
non-linearities are assessed on tasks of regression and classification, using
MNIST as an example. We show a considerable improvement in predictive
log-likelihood and RMSE compared to existing state-of-the-art methods, and
finish by using dropout's uncertainty in deep reinforcement learning.