An Efficient Convolution Core Architecture for Privacy-Preserving Deep Learning

Cloud service for deep learning (DL) has been widely used except for applications involving medical, financial, or other sensitive data due to privacy and security requirements. By employing homomorphic encryption, trained deep convolutional neural networks (CNNs) can be converted to CryptoNets, which is suitable for privacy-preserving DL cloud service. However, the high computation complexity of CryptoNets leads to tremendous implementation challenge. In this paper, to the best of our knowledge, efficient hardware acceleration of CryptoNets is discussed for the first time in open literature. In more detail, without compromise of security and inference accuracy, encryption parameters and modular multiplication algorithm are carefully selected to reduce the computation complexity of polynomial multiplication. Besides, based on the negative wrapped convolution and fast finite impulse filter schemes, an efficient algorithm for convolutions in CryptoNets is developed. Moreover, a dedicated low complexity convolution core architecture for CryptoNets is proposed and implemented with a 90nm CMOS technology. Compared to a well optimized CPU implementation of CryptoNets, this architecture is 11.9× faster while consuming a power of only 537mW.

[1]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[2]  Frederik Vercauteren,et al.  High-Speed Polynomial Multiplication Architecture for Ring-LWE and SHE Cryptosystems , 2015, IEEE Transactions on Circuits and Systems I: Regular Papers.

[3]  Michael Naehrig,et al.  Improved Security for a Ring-Based Fully Homomorphic Encryption Scheme , 2013, IMACC.

[4]  Frederik Vercauteren,et al.  Somewhat Practical Fully Homomorphic Encryption , 2012, IACR Cryptol. ePrint Arch..

[5]  Frederik Vercauteren,et al.  Modular Hardware Architecture for Somewhat Homomorphic Function Evaluation , 2015, CHES.

[6]  Markku Renfors,et al.  The maximum sampling rate of digital filters under hardware speed constraints , 1981 .

[7]  Shen-Fu Hsiao,et al.  Design of high-speed low-power 3-2 counter and 4-2 compressor for fast multipliers , 1998 .

[8]  Ronald L. Rivest,et al.  ON DATA BANKS AND PRIVACY HOMOMORPHISMS , 1978 .

[9]  David Harvey,et al.  Faster arithmetic for number-theoretic transforms , 2012, J. Symb. Comput..

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  Zhongfeng Wang,et al.  Efficient convolution architectures for convolutional neural network , 2016, 2016 8th International Conference on Wireless Communications & Signal Processing (WCSP).

[12]  Arnaud Tisserand,et al.  Hardware/Software Co-Design of an Accelerator for FV Homomorphic Encryption Scheme Using Karatsuba Algorithm , 2018, IEEE Transactions on Computers.

[13]  Michael Naehrig,et al.  CryptoNets: applying neural networks to encrypted data with high throughput and accuracy , 2016, ICML 2016.

[14]  Hao Chen,et al.  Simple Encrypted Arithmetic Library - SEAL v2.1 , 2016, Financial Cryptography Workshops.