Decision Making with Differential Privacy under a Fairness Lens

Many agencies release datasets and statistics about groups of individuals that are used as input to a number of critical decision processes. To conform with privacy and confidentiality requirements, these agencies are often required to release privacypreserving versions of the data. This paper studies the release of differentially private datasets and analyzes their impact on some critical resource allocation tasks under a fairness perspective. The paper shows that, when the decisions take as input differentially private data, the noise added to achieve privacy disproportionately impacts some groups over others. The paper analyzes the reasons for these disproportionate impacts and proposes guidelines to mitigate these effects. The proposed approaches are evaluated on critical decision problems that use differentially private census data.

[1]  John M. Abowd,et al.  The U.S. Census Bureau Adopts Differential Privacy , 2018, KDD.

[2]  Kristina Lerman,et al.  A Survey on Bias and Fairness in Machine Learning , 2019, ACM Comput. Surv..

[3]  Parvez Ahammad,et al.  LinkedIn's Audience Engagements API: A Privacy Preserving Data Analytics System at Scale , 2020, ArXiv.

[4]  Úlfar Erlingsson,et al.  RAPPOR: Randomized Aggregatable Privacy-Preserving Ordinal Response , 2014, CCS.

[5]  Dawn Xiaodong Song,et al.  Towards Practical Differential Privacy for SQL Queries , 2017, Proc. VLDB Endow..

[6]  Michael D. Ekstrand,et al.  Privacy for All: Ensuring Fair and Equitable Privacy Protections , 2018, FAT.

[7]  Aaron Roth,et al.  The Algorithmic Foundations of Differential Privacy , 2014, Found. Trends Theor. Comput. Sci..

[8]  Cynthia Dwork,et al.  Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.

[9]  Pascal Van Hentenryck,et al.  Differentially Private and Fair Deep Learning: A Lagrangian Dual Approach , 2020, AAAI.

[10]  Varun Gupta,et al.  On the Compatibility of Privacy and Fairness , 2019, UMAP.

[11]  Pramod Viswanath,et al.  The Composition Theorem for Differential Privacy , 2013, IEEE Transactions on Information Theory.

[12]  Pascal Van Hentenryck,et al.  Differential Privacy of Hierarchical Census Data: An Optimization Approach , 2019, CP.

[13]  Salil P. Vadhan,et al.  The Complexity of Differential Privacy , 2017, Tutorials on the Foundations of Cryptography.

[14]  Aaron Roth,et al.  Differentially Private Fair Learning , 2018, ICML.

[15]  Pascal Van Hentenryck,et al.  Bias and Variance of Post-processing in Differential Privacy , 2020, AAAI.

[16]  Steffen Rebennack,et al.  Piecewise Linear Function Fitting via Mixed-Integer Linear Programming , 2020, INFORMS J. Comput..

[17]  Ashwin Machanavajjhala,et al.  Fair decision making using privacy-protected data , 2019, FAT*.

[18]  C. Dwork,et al.  Exposed! A Survey of Attacks on Private Data , 2017, Annual Review of Statistics and Its Application.

[19]  Luca Oneto,et al.  Fairness in Machine Learning , 2020, INNSBDDL.

[20]  John M. Abowd,et al.  An Economic Analysis of Privacy Protection and Statistical Accuracy as Social Choices , 2018, American Economic Review.