Machine learning models in health care are often deployed in settings where
it is important to protect patient privacy. In such settings, methods for
differentially private (DP) learning provide a general-purpose approach to
learn models with privacy guarantees. Modern methods for DP learning ensure
privacy through mechanisms that censor information judged as too unique. The
resulting privacy-preserving models, therefore, neglect information from the
tails of a data distribution, resulting in a loss of accuracy that can
disproportionately affect small groups. In this paper, we study the effects of
DP learning in health care. We use state-of-the-art methods for DP learning to
train privacy-preserving models in clinical prediction tasks, including x-ray
classification of images and mortality prediction in time series data. We use
these models to perform a comprehensive empirical investigation of the
tradeoffs between privacy, utility, robustness to dataset shift, and fairness.
Our results highlight lesser-known limitations of methods for DP learning in
health care, models that exhibit steep tradeoffs between privacy and utility,
and models whose predictions are disproportionately influenced by large
demographic groups in the training data. We discuss the costs and benefits of
differentially private learning in health care.