Information-Theoretic Privacy in Federated Submodel learning

We consider information-theoretic privacy in federated submodel learning, where a global server has multiple submodels. Compared to the privacy considered in the conventional federated submodel learning where secure aggregation is adopted for ensuring privacy, information-theoretic privacy provides the stronger protection on submodel selection by the local machine. We propose an achievable scheme that partially adopts the conventional private information retrieval (PIR) scheme that achieves the minimum amount of download. With respect to computation and communication overhead, we compare the achievable scheme with a naive approach for federated submodel learning with information-theoretic privacy.

[1]  Shaojie Tang,et al.  Secure Federated Submodel Learning , 2019, ArXiv.

[2]  Eyal Kushilevitz,et al.  Private information retrieval , 1998, JACM.

[3]  Hua Sun,et al.  The Capacity of Private Information Retrieval , 2017, IEEE Transactions on Information Theory.

[4]  Sarvar Patel,et al.  Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..