Understanding mean-field effects of large-population user data obfuscation in machine learning

Recently, data-based services have bloomed, because of the prevalence of online tracking, wearable computing, and the Internet of Things. However, due to privacy concerns, users may use tools to obfuscate their data, rendering these services less useful. This conflict places the service's desire for accuracy, and the users' desire for both accuracy and privacy, in contention. We propose a game-theoretic model for this conflict. By promising bounds on how much data gets collected and sold, services can incentivize users to report their data truthfully. We model this incentivization process as a Stackelberg game with the service provider as the leader. The users react to privacy promises as Stackelberg followers and interact with each other by playing a mean-field game. By representing the users as an interval on the continuum, we reduce the computational complexity of finding Nash and Stackelberg equilibriums.

[1]  Finn Brunton,et al.  Obfuscation: A User's Guide for Privacy and Protest , 2015 .

[2]  Quanyan Zhu,et al.  A mean-field stackelberg game approach for obfuscation adoption in empirical risk minimization , 2017, 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP).

[3]  Quanyan Zhu,et al.  Risk-Sensitive Mean-Field Games , 2012, IEEE Transactions on Automatic Control.

[4]  Helen Nissenbaum,et al.  Trackmenot: Resisting Surveillance in Web Search , 2015 .

[5]  Quanyan Zhu,et al.  Hybrid risk-sensitive mean-field stochastic differential games with application to molecular biology , 2011, IEEE Conference on Decision and Control and European Control Conference.

[6]  Pawlick Jeffrey,et al.  A Stackelberg game perspective on the conflict between machine learning and data obfuscation , 2016 .

[7]  Cynthia Dwork,et al.  Differential Privacy , 2006, ICALP.

[8]  Anand D. Sarwate,et al.  Differentially Private Empirical Risk Minimization , 2009, J. Mach. Learn. Res..