These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Despite the dominant role of deep models in machine learning, limitations
persist, including overconfident predictions, susceptibility to adversarial
attacks, and underestimation of variability in predictions. The Bayesian
paradigm provides a natural framework to overcome such issues and has become
the gold standard for uncertainty estimation with deep models, also providing
improved accuracy and a framework for tuning critical hyperparameters. However,
exact Bayesian inference is challenging, typically involving variational
algorithms that impose strong independence and distributional assumptions.
Moreover, existing methods are sensitive to the architectural choice of the
network. We address these issues by constructing a relaxed version of the
standard feed-forward rectified neural network, and employing Polya-Gamma data
augmentation tricks to render a conditionally linear and Gaussian model.
Additionally, we use sparsity-promoting priors on the weights of the neural
network for data-driven architectural design. To approximate the posterior, we
derive a variational inference algorithm that avoids distributional assumptions
and independence across layers and is a faster alternative to the usual Markov
Chain Monte Carlo schemes.