These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
This paper describes a differentially private post-processing algorithm for
learning fair regressors satisfying statistical parity, addressing privacy
concerns of machine learning models trained on sensitive data, as well as
fairness concerns of their potential to propagate historical biases. Our
algorithm can be applied to post-process any given regressor to improve
fairness by remapping its outputs. It consists of three steps: first, the
output distributions are estimated privately via histogram density estimation
and the Laplace mechanism, then their Wasserstein barycenter is computed, and
the optimal transports to the barycenter are used for post-processing to
satisfy fairness. We analyze the sample complexity of our algorithm and provide
fairness guarantee, revealing a trade-off between the statistical bias and
variance induced from the choice of the number of bins in the histogram, in
which using less bins always favors fairness at the expense of error.
External Datasets
Communities & Crime
Law School
References
Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security
Deep Learning with Differential Privacy
Martín Abadi, Andy Chu, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Kunal Talwar, Li Zhang
Published: 2016
Proceedings of the 32nd International Conference on Algorithmic Learning Theory
On the Sample Complexity of Privately Learning Unbounded High-Dimensional Gaussians
Ishaq Aden-Ali, Hassan Ashtiani, Gautam Kamath
Published: 2021
International Conference on Machine Learning
A reductions approach to fair classification
Alekh Agarwal, Alina Beygelzimer, Miroslav Dudik, John Langford, Hanna Wallach
Published: 2018
Proceedings of the 36th International Conference on Machine Learning
Fair Regression: Quantitative Definitions and Reduction-Based Algorithms
Alekh Agarwal, Miroslav Dudík, Zhiwei Steven Wu
Published: 2019
IJCAI 2020 Workshop on AI for Social Good
Trade-Offs between Fairness and Privacy in Machine Learning
Calibrating noise to sensitivity in private data analysis
Cynthia Dwork, Frank McSherry, Kobbi Nissim, Adam Smith
Published: 2006
Proceedings of the 3rd Innovations in Theoretical Computer Science Conference
Fairness Through Awareness
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, Richard Zemel
Published: 2012
ACM Computing Surveys
Decision Tree Classification with Differential Privacy: A Survey
Sam Fletcher, Md. Zahidul Islam
Published: 2019
Proceedings of the Forty-Second ACM Symposium on Theory of Computing
On the Geometry of Differential Privacy
Moritz Hardt, Kunal Talwar
Published: 2010
arxiv
Cited by 1
NIPS
Equality of Opportunity in Supervised Learning
Moritz Hardt, Eric Price, Nathan Srebro
Published: 10.8.2016
We propose a criterion for discrimination against a specified sensitive
attribute in supervised learning, where the goal is to predict some target
based on available features. Assuming data about the predictor, target, and
membership in the protected group are available, we show how to optimally
adjust any learned predictor so as to remove discrimination according to our
definition. Our framework also improves incentives by shifting the cost of poor
classification from disadvantaged groups to the decision maker, who can respond
by improving the classification accuracy.
In line with other studies, our notion is oblivious: it depends only on the
joint statistics of the predictor, the target and the protected attribute, but
not on interpretation of individualfeatures. We study the inherent limits of
defining and identifying biases based on such oblivious measures, outlining
what can and cannot be inferred from different oblivious tests.
We illustrate our notion using a case study of FICO credit scores.
Proceedings of the 37th International Conference on Machine Learning
Fair Learning with Private Demographic Data
Hussein Mozannar, Mesrob I. Ohannessian, Nathan Srebro
Published: 2020
The Tenth International Conference on Learning Representations
Hyperparameter Tuning with Renyi Differential Privacy
Nicolas Papernot, Thomas Steinke
Published: 2022
arxiv
Cited by 1
International Conference on Learning Representations (ICLR)
Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data
Nicolas Papernot, Martín Abadi, Úlfar Erlingsson, Ian Goodfellow, Kunal Talwar
Published: 10.19.2016
Some machine learning applications involve training data that is sensitive,
such as the medical histories of patients in a clinical trial. A model may
inadvertently and implicitly store some of its training data; careful analysis
of the model may therefore reveal sensitive information.
To address this problem, we demonstrate a generally applicable approach to
providing strong privacy guarantees for training data: Private Aggregation of
Teacher Ensembles (PATE). The approach combines, in a black-box fashion,
multiple models trained with disjoint datasets, such as records from different
subsets of users. Because they rely directly on sensitive data, these models
are not published, but instead used as "teachers" for a "student" model. The
student learns to predict an output chosen by noisy voting among all of the
teachers, and cannot directly access an individual teacher or the underlying
data or parameters. The student's privacy properties can be understood both
intuitively (since no single teacher and thus no single dataset dictates the
student's training) and formally, in terms of differential privacy. These
properties hold even if an adversary can not only query the student but also
inspect its internal workings.
Compared with previous work, the approach imposes only weak assumptions on
how teachers are trained: it applies to any model, including non-convex models
like DNNs. We achieve state-of-the-art privacy/utility trade-offs on MNIST and
SVHN thanks to an improved privacy analysis and semi-supervised learning.