Minimum precision requirements for the SVM-SGD learning algorithm

It is well-known that the precision of data, weight vector, and internal representations employed in learning systems directly impacts their energy, throughput, and latency. The precision requirements for the training algorithm are also important for systems that learn on-the-fly. In this paper, we present analytical lower bounds on the precision requirements for the commonly employed stochastic gradient descent (SGD) on-line learning algorithm in the specific context of a support vector machine (SVM). These bounds are obtained subject to desired system performance. These bounds are validated using the UCI breast cancer dataset. Additionally, the impact of these precisions on the energy consumption of a fixed-point SVM with on-line training is studied. Simulation results in 45 nm CMOS process show that operating at the minimum precision as dictated by our bounds improves energy consumption by a factor of 5.3× as compared to conventional precision assignments with no observable loss in accuracy.

[1]  Sachin S. Talathi,et al.  Fixed Point Quantization of Deep Convolutional Networks , 2015, ICML.

[2]  Léon Bottou,et al.  Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.

[3]  Charbel Sakr,et al.  Understanding the Energy and Precision Requirements for Online Learning , 2016, ArXiv.

[4]  Ali Farhadi,et al.  XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.

[5]  Naresh R. Shanbhag,et al.  Finite-precision analysis of the pipelined strength-reduced adaptive filter , 1998, IEEE Trans. Signal Process..

[6]  Yoshua Bengio,et al.  BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 , 2016, ArXiv.

[7]  Sachin S. Talathi,et al.  Overcoming Challenges in Fixed Point Training of Deep Convolutional Networks , 2016, ArXiv.

[8]  Keshab K. Parhi,et al.  VLSI digital signal processing systems , 1999 .

[9]  Yoshua Bengio,et al.  BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.

[10]  Naresh R. Shanbhag,et al.  Reducing Energy at the Minimum Energy Operating Point Via Statistical Error Compensation , 2014, IEEE Transactions on Very Large Scale Integration (VLSI) Systems.

[11]  Pritish Narayanan,et al.  Deep Learning with Limited Numerical Precision , 2015, ICML.

[12]  Ran El-Yaniv,et al.  Binarized Neural Networks , 2016, NIPS.

[13]  Paris Smaragdis,et al.  Bitwise Neural Networks , 2016, ArXiv.

[14]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.