A Stackelberg game perspective on the conflict between machine learning and data obfuscation

Data is the new oil; this refrain is repeated extensively in the age of internet tracking, machine learning, and data analytics. As data collection becomes more personal and pervasive, however, public pressure is mounting for privacy protection. In this atmosphere, developers have created applications to add noise to user attributes visible to tracking algorithms. This creates a strategic interaction between trackers and users when incentives to maintain privacy and improve accuracy are misaligned. In this paper, we conceptualize this conflict through an N + 1-player, augmented Stackelberg game. First a machine learner declares a privacy protection level, and then users respond by choosing their own perturbation amounts. We use the general frameworks of differential privacy and empirical risk minimization to quantify the utility components due to privacy and accuracy, respectively. In equilibrium, each user perturbs her data independently, which leads to a high net loss in accuracy. To remedy this scenario, we show that the learner improves his utility by proactively perturbing the data himself. While other work in this area has studied privacy markets and mechanism design for truthful reporting of user information, we take a different viewpoint by considering both user and learnerperturbation.

[1]  Nathan Srebro,et al.  Fast Rates for Regularized Objectives , 2008, NIPS.

[2]  Jens Grossklags,et al.  A Game-Theoretic Study on Non-monetary Incentives in Data Analytics Projects with Privacy Implications , 2015, 2015 IEEE 28th Computer Security Foundations Symposium.

[3]  John C. Mitchell,et al.  Third-Party Web Tracking: Policy and Technology , 2012, 2012 IEEE Symposium on Security and Privacy.

[4]  Carmela Troncoso,et al.  Protecting location privacy: optimal strategy against localization attacks , 2012, CCS.

[5]  Wayne H. Wolf,et al.  Cyber-physical Systems , 2009, Computer.

[6]  Ting Yu,et al.  Empirical privacy and empirical utility of anonymized data , 2013, 2013 IEEE 29th International Conference on Data Engineering Workshops (ICDEW).

[7]  H. Stackelberg,et al.  Marktform und Gleichgewicht , 1935 .

[8]  R.S. Parker,et al.  A model-based algorithm for blood glucose control in Type I diabetic patients , 1999, IEEE Transactions on Biomedical Engineering.

[9]  Latanya Sweeney,et al.  k-Anonymity: A Model for Protecting Privacy , 2002, Int. J. Uncertain. Fuzziness Knowl. Based Syst..

[10]  ASHWIN MACHANAVAJJHALA,et al.  L-diversity: privacy beyond k-anonymity , 2006, 22nd International Conference on Data Engineering (ICDE'06).

[11]  Melanie Swan,et al.  Sensor Mania! The Internet of Things, Wearable Computing, Objective Metrics, and the Quantified Self 2.0 , 2012, J. Sens. Actuator Networks.

[12]  Aaron Roth,et al.  Selling privacy at auction , 2015, Games Econ. Behav..

[13]  Anand D. Sarwate,et al.  Differentially Private Empirical Risk Minimization , 2009, J. Mach. Learn. Res..

[14]  Scott R. Peppet Regulating the Internet of Things: First Steps Toward Managing Discrimination, Privacy, Security & Consent , 2014 .

[15]  Cynthia Dwork,et al.  Differential Privacy , 2006, ICALP.

[16]  Moni Naor,et al.  On the Difficulties of Disclosure Prevention in Statistical Databases or The Case for Differential Privacy , 2010, J. Priv. Confidentiality.

[17]  Aaron Roth,et al.  The Algorithmic Foundations of Differential Privacy , 2014, Found. Trends Theor. Comput. Sci..

[18]  Jens Grossklags,et al.  A Short Paper on the Incentives to Share Private Information for Population Estimates , 2015, Financial Cryptography.

[19]  Rico Neumann,et al.  Obfuscation: A user’s guide for privacy and protest , 2017, New Media & Society.

[20]  Ian R. Kerr,et al.  Lessons from the Identity Trail: Anonymity, Privacy and Identity in a Networked Society , 2009 .

[21]  Ninghui Li,et al.  Membership privacy: a unifying framework for privacy definitions , 2013, CCS.

[22]  David Xiao,et al.  Is privacy compatible with truthfulness? , 2013, ITCS '13.

[23]  Helen Nissenbaum,et al.  Trackmenot: Resisting Surveillance in Web Search , 2015 .

[24]  Romit Roy Choudhury,et al.  Hiding stars with fireworks: location privacy through camouflage , 2009, MobiCom '09.