Newton-type Methods for Minimax Optimization
暂无分享,去创建一个
[1] T. Rutherford,et al. Nonlinear Programming , 2021, Mathematical Programming Methods for Geographers and Planners.
[2] Tanner Fiez,et al. Implicit Learning Dynamics in Stackelberg Games: Equilibria Characterization, Convergence Analysis, and Empirical Study , 2020, ICML.
[3] Yaoliang Yu,et al. Optimality and Stability in Non-Convex Smooth Games , 2020, J. Mach. Learn. Res..
[4] Yaoliang Yu,et al. Optimality and Stability in Non-Convex-Non-Concave Min-Max Optimization , 2020, ArXiv.
[5] Asuman Ozdaglar,et al. GANs May Have No Nash Equilibria , 2020, ICML 2020.
[6] Junyu Zhang,et al. On Lower Iteration Complexity Bounds for the Saddle Point Problems , 2019 .
[7] Xingshi He,et al. Introduction to Optimization , 2015, Applied Evolutionary Algorithms for Engineers Using Python.
[8] Jimmy Ba,et al. On Solving Minimax Optimization Locally: A Follow-the-Ridge Approach , 2019, ICLR.
[9] J. Malick,et al. On the convergence of single-call stochastic extra-gradient methods , 2019, NeurIPS.
[10] Yaoliang Yu,et al. Convergence of Gradient Methods on Bilinear Zero-Sum Games , 2019, ICLR.
[11] Ioannis Mitliagkas,et al. Linear Lower Bounds and Conditioning of Differentiable Games , 2019, ICML.
[12] Ioannis Mitliagkas,et al. Lower Bounds and Conditioning of Differentiable Games , 2019, ArXiv.
[13] Lillian J. Ratliff,et al. Convergence of Learning Dynamics in Stackelberg Games , 2019, ArXiv.
[14] Michael I. Jordan,et al. What is Local Optimality in Nonconvex-Nonconcave Minimax Optimization? , 2019, ICML.
[15] Michael I. Jordan,et al. Minmax Optimization: Stable Limit Points of Gradient Descent Ascent are Locally Optimal , 2019, ArXiv.
[16] Mehryar Mohri,et al. Agnostic Federated Learning , 2019, ICML.
[17] Stefano Ermon,et al. Learning Controllable Fair Representations , 2018, AISTATS.
[18] Yangyang Xu,et al. Lower complexity bounds of first-order methods for convex-concave bilinear saddle-point problems , 2018, Math. Program..
[19] Chuan-Sheng Foo,et al. Optimistic mirror descent in saddle-point problems: Going the extra (gradient) mile , 2018, ICLR.
[20] Stephen J. Wright,et al. A Newton-CG algorithm with complexity guarantees for smooth unconstrained optimization , 2018, Mathematical Programming.
[21] Thore Graepel,et al. The Mechanics of n-Player Differentiable Games , 2018, ICML.
[22] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[23] Sashank J. Reddi,et al. On the Convergence of Adam and Beyond , 2018, ICLR.
[24] Le Song,et al. SBEED: Convergent Reinforcement Learning with Nonlinear Function Approximation , 2017, ICML.
[25] Shimon Whiteson,et al. Learning with Opponent-Learning Awareness , 2017, AAMAS.
[26] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[27] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[28] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[29] J. Zico Kolter,et al. Gradient descent GAN optimization is locally stable , 2017, NIPS.
[30] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[31] Lihong Li,et al. Stochastic Variance Reduction Methods for Policy Evaluation , 2017, ICML.
[32] Yann LeCun,et al. Singularity of the Hessian in Deep Learning , 2016, ArXiv.
[33] David Pfau,et al. Unrolled Generative Adversarial Networks , 2016, ICLR.
[34] François Laviolette,et al. Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..
[35] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[36] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[37] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[38] Stephen P. Boyd,et al. Convex Optimization , 2004, IEEE Transactions on Automatic Control.
[39] V. Borkar. Stochastic Approximation: A Dynamical Systems Viewpoint , 2008 .
[40] Robert Leonard. Theory of Games and Economic Behavior , 2006 .
[41] Carl D. Meyer,et al. Matrix Analysis and Applied Linear Algebra , 2000 .
[42] Barak A. Pearlmutter. Fast Exact Multiplication by the Hessian , 1994, Neural Computation.
[43] P. Werbos. Backpropagation: past and future , 1988, ICNN.
[44] L. Popov. A modification of the Arrow-Hurwicz method for search of saddle points , 1980 .
[45] J. Nocedal. Updating Quasi-Newton Matrices With Limited Storage , 1980 .
[46] B. V. Dean,et al. Studies in Linear and Non-Linear Programming. , 1959 .
[47] J. Neumann,et al. Theory of games and economic behavior , 1945, 100 Years of Math Milestones.
[48] Asuman Ozdaglar,et al. Do GANs always have Nash equilibria? , 2020, ICML.
[49] Jinsung Yoon,et al. GENERATIVE ADVERSARIAL NETS , 2018 .
[50] John C. Duchi,et al. Certifiable Distributional Robustness with Principled Adversarial Training , 2017, ArXiv.
[51] Geoffrey E. Hinton,et al. Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks , 2006 .
[52] Arkadi Nemirovski,et al. Prox-Method with Rate of Convergence O(1/t) for Variational Inequalities with Lipschitz Continuous Monotone Operators and Smooth Convex-Concave Saddle Point Problems , 2004, SIAM J. Optim..
[53] E. Chong,et al. Introduction to optimization , 1987 .
[54] G. M. Korpelevich. The extragradient method for finding saddle points and other problems , 1976 .
[55] Yu. G. Evtushenko,et al. Some local properties of minimax problems , 1974 .
[56] Yu. G. Evtushenko,et al. Iterative methods for solving minimax problems , 1974 .
[57] Boris Polyak. Some methods of speeding up the convergence of iteration methods , 1964 .
[58] J. Nash. Equilibrium Points in N-Person Games. , 1950, Proceedings of the National Academy of Sciences of the United States of America.