Predictive privacy: Collective data protection in the context of artificial intelligence and big data

Big data and artificial intelligence pose a new challenge for data protection as these techniques allow predictions to be made about third parties based on the anonymous data of many people. Examples of predicted information include purchasing power, gender, age, health, sexual orientation, ethnicity, etc. The basis for such applications of “predictive analytics” is the comparison between behavioral data (e.g. usage, tracking, or activity data) of the individual in question and the potentially anonymously processed data of many others using machine learning models or simpler statistical methods. The article starts by noting that predictive analytics has a significant potential to be abused, which manifests itself in the form of social inequality, discrimination, and exclusion. These potentials are not regulated by current data protection law in the EU; indeed, the use of anonymized mass data takes place in a largely unregulated space. Under the term “predictive privacy,” a data protection approach is presented that counters the risks of abuse of predictive analytics. A person's predictive privacy is violated when personal information about them is predicted without their knowledge and against their will based on the data of many other people. Predictive privacy is then formulated as a protected good and improvements to data protection with regard to the regulation of predictive analytics are proposed. Finally, the article points out that the goal of data protection in the context of predictive analytics is the regulation of “prediction power,” which is a new manifestation of informational power asymmetry between platform companies and society.

[1]  J. Muldoon Data-owning democracy or digital socialism? , 2022, Critical Review of International Social and Political Philosophy.

[2]  Roberta Fischli Data-owning democracy: Citizen empowerment through data ownership , 2022, European Journal of Political Theory.

[3]  Jacob Leon Kröger,et al.  The myth of individual control: Mapping the limitations of privacy self-management , 2021, SSRN Electronic Journal.

[4]  Rainer Mühlhoff Automatisierte Ungleichheit , 2020 .

[5]  Patrick Skeba,et al.  Informational Friction as a Lens for Studying Algorithmic Aspects of Privacy , 2020, Proc. ACM Hum. Comput. Interact..

[6]  Rainer Mühlhoff Predictive privacy: towards an applied ethics of data analytics , 2020, Ethics and Information Technology.

[7]  Jordan M. Blanke Protection for ‘Inferences Drawn’: A Comparison Between the General Data Protection Regulation and the California Consumer Privacy Act , 2020, Global Privacy Law Review.

[8]  Sharath Chandra Guntuku,et al.  Evaluating the predictability of medical conditions from social media posts , 2019, PloS one.

[9]  Shikha Verma,et al.  Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy , 2019, Vikalpa: The Journal for Decision Makers.

[10]  M. Christen,et al.  Two Concepts of Group Privacy , 2019, Philosophy & Technology.

[11]  Filippo Lorè [Evaluating the impact of regulation 2016/679 of the European Parliament]. , 2019, Giornale Italiano di Nefrologia.

[12]  Tal Z. Zarsky Privacy and Manipulation in the Digital Age , 2019, Theoretical Inquiries in Law.

[13]  Sandra Wachter Data protection in the age of big data , 2019, Nature Electronics.

[14]  O. Ben-shahar,et al.  Data Pollution , 2019, Journal of Legal Analysis.

[15]  H. Nissenbaum,et al.  Online Manipulation: Hidden Influences in a Digital World , 2018 .

[16]  Rainer Mühlhoff Digitale Entmündigung und User Experience Design. Wie digitale Geräte uns nudgen, tracken und zur Unwissenheit erziehen , 2018 .

[17]  Hannah Lebovits Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor , 2018, Public Integrity.

[18]  Sandra Wachter,et al.  A Right to Reasonable Inferences: Re-Thinking Data Protection Law in the Age of Big Data and AI , 2018 .

[19]  D. Fitch,et al.  Review of "Algorithms of oppression: how search engines reinforce racism," by Noble, S. U. (2018). New York, New York: NYU Press. , 2018, CDQR.

[20]  S. Noble Algorithms of Oppression: How Search Engines Reinforce Racism , 2018 .

[21]  Mihaela Popescu,et al.  Big data analytics and the limits of privacy self-management , 2017, New Media Soc..

[22]  Brent Mittelstadt,et al.  From Individual to Group Privacy in Big Data Analytics , 2017 .

[23]  Paula Helm,et al.  Group Privacy in Times of Big Data. A Literature Review , 2016 .

[24]  Linnet Taylor,et al.  The ethics of big data as a public good: which public? Whose good? , 2016, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[25]  Ian Goodfellow,et al.  Deep Learning with Differential Privacy , 2016, CCS.

[26]  Karen Yeung,et al.  ‘Hypernudge’: Big Data as a mode of regulation by design , 2016, The Social Power of Algorithms.

[27]  Alessandro Mantelero,et al.  Personal data for decisional purposes in the age of analytics: From an individual to a collective dimension of data protection , 2016, Comput. Law Secur. Rev..

[28]  Lorraine Code The Myth of the Individual , 2016, The American journal of bioethics : AJOB.

[29]  L. Floridi Open Data, Data Protection, and Group Privacy , 2014, Philosophy & Technology.

[30]  Robert H. Sloan,et al.  Beyond Notice and Choice: Privacy, Norms, and Consent , 2013 .

[31]  T. Graepel,et al.  Private traits and attributes are predictable from digital records of human behavior , 2013, Proceedings of the National Academy of Sciences.

[32]  Eran Halperin,et al.  Identifying Personal Genomes by Surname Inference , 2013, Science.

[33]  H. Nissenbaum A Contextual Approach to Privacy Online , 2011, Daedalus.

[34]  Paul Ohm Broken Promises of Privacy: Responding to the Surprising Failure of Anonymization , 2009 .

[35]  Vitaly Shmatikov,et al.  Robust De-anonymization of Large Sparse Datasets , 2008, 2008 IEEE Symposium on Security and Privacy (sp 2008).

[36]  Serge Gutwirth,et al.  Profiling the European Citizen, Cross-Disciplinary Perspectives , 2008 .

[37]  Cynthia Dwork,et al.  Differential Privacy , 2006, ICALP.

[38]  L. Sweeney,et al.  k-Anonymity: A Model for Protecting Privacy , 2002, Int. J. Uncertain. Fuzziness Knowl. Based Syst..

[39]  Priscilla M. Regan Privacy as a Common Good in the Digital World , 2002 .

[40]  Anton Vedder,et al.  KDD: The challenge to individualism , 1999, Ethics and Information Technology.

[41]  Joseph Gray Jackson,et al.  Privacy and Freedom , 1968 .

[42]  L. Floridi,et al.  Book review : Group Privacy : New Challenges of Data Technologies , 2017 .

[43]  Dear Mr Sotiropoulos ARTICLE 29 Data Protection Working Party , 2013 .

[44]  Kai von Lewinski,et al.  Geschichte des Datenschutzrechts von 1600 bis 1977 , 2009 .

[45]  Herman T. Tavani,et al.  PHILOSOPHICAL THEORIES OF PRIVACY: IMPLICATIONS FOR AN ADEQUATE ONLINE PRIVACY POLICY , 2007 .