Application of Quantum Annealing to Training of Deep Neural Networks

In Deep Learning, a well-known approach for training a Deep Neural Network starts by training a generative Deep Belief Network model, typically using Contrastive Divergence (CD), then fine-tuning the weights using backpropagation or other discriminative techniques. However, the generative training can be time-consuming due to the slow mixing of Gibbs sampling. We investigated an alternative approach that estimates model expectations of Restricted Boltzmann Machines using samples from a D-Wave quantum annealing machine. We tested this method on a coarse-grained version of the MNIST data set. In our tests we found that the quantum sampling-based training approach achieves comparable or better accuracy with significantly fewer iterations of generative training than conventional CD-based training. Further investigation is needed to determine whether similar improvements can be achieved for other data sets, and to what extent these improvements can be attributed to quantum effects.

[1]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[2]  M. W. Johnson,et al.  Quantum annealing with manufactured spins , 2011, Nature.

[3]  S. Knysh,et al.  Quantum Optimization of Fully-Connected Spin Glasses , 2014, 1406.7553.

[4]  B. Chakrabarti,et al.  Multivariable optimization: Quantum annealing and computation , 2014, The European Physical Journal Special Topics.

[5]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[6]  Misha Denil Toward the Implementation of a Quantum RBM , 2011 .

[7]  Noboru Murata,et al.  Nonparametric Weight Initialization of Neural Networks via Integral Representation , 2013, ArXiv.

[8]  Tommy W. S. Chow,et al.  A weight initialization method for improving training speed in feedforward neural network , 2000, Neurocomputing.

[9]  Klaus-Robert Müller,et al.  Deep Boltzmann Machines and the Centering Trick , 2012, Neural Networks: Tricks of the Trade.

[10]  Daniel A. Lidar,et al.  Defining and detecting quantum speedup , 2014, Science.

[11]  Geoffrey E. Hinton Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.

[12]  Nando de Freitas,et al.  Toward the Implementation of a Quantum RBM , 2011 .

[13]  Ashish Kapoor,et al.  Quantum deep learning , 2014, Quantum Inf. Comput..

[14]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[15]  Yoshua Bengio,et al.  On the Challenges of Physical Implementations of RBMs , 2013, AAAI.

[16]  M. Benedetti,et al.  Estimation of effective temperatures in quantum annealers for sampling applications: A case study with possible applications in deep learning , 2015, 1510.07611.

[17]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[18]  Daniel A. Lidar,et al.  Experimental signature of programmable quantum annealing , 2012, Nature Communications.

[19]  J. F. Shepanski Fast learning in artificial neural systems: multilayer perceptron training using optimal estimation , 1988, IEEE 1988 International Conference on Neural Networks.

[20]  Tapani Raiko,et al.  NIPS 2011 Workshop on Deep Learning and Unsupervised Feature Learning , 2011 .

[21]  R. Flinn The challenges. , 1979, Delaware medical journal.