Deep Learning Meets Game Theory: Bregman-Based Algorithms for Interactive Deep Generative Adversarial Networks

This paper presents an interplay between deep learning and game theory. It models basic deep learning tasks as strategic games. Then, distributionally robust games and their relationship with deep generative adversarial networks (GANs) are presented. To achieve a higher order convergence rate without using a second derivative of the objective function, a Bregman discrepancy is used to construct a speed-up deep learning. Each player has a continuous action space which corresponds to weight space and aims to learn his/her optimal strategy. The convergence rate of the proposed deep learning algorithm is derived using a mean estimate. Experiments are carried out on a real dataset in both shallow and deep GANs. Both qualitative and quantitative evaluation results show that the generative model trained by the Bregman deep learning algorithm can speed up the state-of-the-art performance.

[1]  M. Sion On general minimax theorems , 1958 .

[2]  H. Robbins A Stochastic Approximation Method , 1951 .

[3]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[4]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[5]  Matthew D. Zeiler ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.

[6]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[7]  S. Linnainmaa Taylor expansion of the accumulated rounding error , 1976 .

[8]  Hamidou Tembine,et al.  Distributionally Robust Games: f-Divergence and Learning , 2017, VALUETOOLS.

[9]  David Balduzzi,et al.  Deep Online Convex Optimization with Gated Games , 2016, ArXiv.

[10]  L. Armijo Minimization of functions having Lipschitz continuous first partial derivatives. , 1966 .

[11]  Andre Wibisono,et al.  A variational perspective on accelerated methods in optimization , 2016, Proceedings of the National Academy of Sciences.

[12]  Pradeep Dubey,et al.  Strategic complements and substitutes, and potential games , 2006, Games Econ. Behav..

[13]  Dale Schuurmans,et al.  Deep Learning Games , 2016, NIPS.

[14]  Boris Polyak Some methods of speeding up the convergence of iteration methods , 1964 .

[15]  Hamidou Tembine Distributed Strategic Learning for Wireless Engineers , 2017 .

[16]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[17]  Pierre Priouret,et al.  Adaptive Algorithms and Stochastic Approximations , 1990, Applications of Mathematics.

[18]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[19]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[20]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[21]  David Balduzzi,et al.  Cortical prediction markets , 2014, AAMAS.

[22]  Alekseĭ Grigorʹevich Ivakhnenko,et al.  CYBERNETIC PREDICTING DEVICES , 1966 .