暂无分享,去创建一个
Michael I. Jordan | Lillian J. Ratliff | Eric Mazumdar | S. Shankar Sastry | Eric V. Mazumdar | S. Sastry | L. Ratliff
[1] R. E. Kalman,et al. Contributions to the Theory of Optimal Control , 1960 .
[2] D. Lukes,et al. A GLOBAL THEORY FOR LINEAR QUADRATIC DIFFERENTIAL GAMES. , 1971 .
[3] T. Başar. On the uniqueness of the Nash solution in Linear-Quadratic differential Games , 1976 .
[4] T. Başar,et al. Dynamic Noncooperative Game Theory , 1982 .
[5] R. Abraham,et al. Manifolds, Tensor Analysis, and Applications , 1983 .
[6] M. Shub. Global Stability of Dynamical Systems , 1986 .
[7] T.-Y. Li,et al. Lyapunov Iterations for Solving Coupled Algebraic Riccati Equations of Nash Differential Games and Algebraic Riccati Equations of Zero-Sum Games , 1995 .
[8] T. Başar. Contributions to the Theory of Optimal Control , 2001 .
[9] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[10] S. Smale. Differentiable dynamical systems , 1967 .
[11] Cars H. Hommes,et al. Multiple equilibria and limit cycles in evolutionary games with Logit Dynamics , 2012, Games Econ. Behav..
[12] S. Shankar Sastry,et al. Characterization and computation of local Nash equilibria in continuous games , 2013, 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[13] Corrado Possieri,et al. An algebraic geometry approach for the computation of all linear feedback Nash equilibria in LQ differential games , 2015, 2015 54th IEEE Conference on Decision and Control (CDC).
[14] Yi Wu,et al. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.
[15] David Silver,et al. A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning , 2017, NIPS.
[16] Constantinos Daskalakis,et al. Training GANs with Optimism , 2017, ICLR.
[17] Michael H. Bowling,et al. Actor-Critic Policy Optimization in Partially Observable Multiagent Environments , 2018, NeurIPS.
[18] Sham M. Kakade,et al. Global Convergence of Policy Gradient Methods for the Linear Quadratic Regulator , 2018, ICML.
[19] Jakub W. Pachocki,et al. Emergent Complexity via Multi-Agent Competition , 2017, ICLR.
[20] Christos H. Papadimitriou,et al. Cycles in adversarial regularized learning , 2017, SODA.
[21] Tamer Basar,et al. Policy Optimization Provably Converges to Nash Equilibria in Zero-Sum Linear Quadratic Games , 2019, NeurIPS.
[22] Martin J. Wainwright,et al. Derivative-Free Methods for Policy Optimization: Guarantees for Linear Quadratic Systems , 2018, AISTATS.
[23] S. Shankar Sastry,et al. On Finding Local Nash Equilibria (and Only Local Nash Equilibria) in Zero-Sum Games , 2019, 1901.00838.
[24] Guy Lever,et al. Human-level performance in 3D multiplayer games with population-based reinforcement learning , 2018, Science.
[25] S. Shankar Sastry,et al. On Gradient-Based Learning in Continuous Games , 2018, SIAM J. Math. Data Sci..
[26] Nikolai Matni,et al. On the Sample Complexity of the Linear Quadratic Regulator , 2017, Foundations of Computational Mathematics.