Hypothesis testing under maximal leakage privacy constraints

The problem of publishing privacy-guaranteed data for hypothesis testing is studied using the maximal leakage (ML) as a metric for privacy and the type-II error exponent as the utility metric. The optimal mechanism (random mapping) that maximizes utility for a bounded leakage guarantee is determined for the entire leakage range for binary datasets. For non-binary datasets, approximations in the high privacy and high utility regimes are developed. The results show that, for any desired leakage level, maximizing utility forces the ML privacy mechanism to reveal partial to complete knowledge about a subset of the source alphabet. The results developed on maximizing a convex function over a polytope may also of an independent interest.

[1]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[2]  Lizhong Zheng,et al.  Euclidean Information Theory , 2008, 2008 IEEE International Zurich Seminar on Communications.

[3]  Mário S. Alvim,et al.  Measuring Information Leakage Using Generalized Gain Functions , 2012, 2012 IEEE 25th Computer Security Foundations Symposium.

[4]  Shao-Lun Huang,et al.  Euclidean Information Theory of Networks , 2015, IEEE Transactions on Information Theory.

[5]  Vincent Y. F. Tan,et al.  Hypothesis testing in the high privacy limit , 2016, 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[6]  Sudeep Kamath,et al.  An operational measure of information leakage , 2016, 2016 Annual Conference on Information Science and Systems (CISS).