Fast Gradient Descent for Drifting Least Squares Regression, with Application to Bandits
暂无分享,去创建一个
Nathaniel Korda | A. PrashanthL. | Rémi Munos | R. Munos | N. Korda | A. PrashanthL. | Prashanth L.A.
[1] Rémi Munos,et al. Online gradient descent for least squares regression: Non-asymptotic bounds and application to bandits , 2013, ArXiv.
[2] S. Menozzi,et al. Concentration bounds for stochastic approximations , 2012, 1204.3730.
[3] Eric Moulines,et al. Non-Asymptotic Analysis of Stochastic Approximation Algorithms for Machine Learning , 2011, NIPS.
[4] Thomas P. Hayes,et al. Stochastic Linear Optimization under Bandit Feedback , 2008, COLT.
[5] Martin Zinkevich,et al. Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.
[6] Wei Chu,et al. A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.
[7] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[8] M. Fathi,et al. Transport-Entropy inequalities and deviation estimates for stochastic approximation schemes , 2013, 1301.7740.
[9] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[10] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[11] John Darzentas,et al. Problem Complexity and Method Efficiency in Optimization , 1983 .
[12] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[13] John N. Tsitsiklis,et al. Linearly Parameterized Bandits , 2008, Math. Oper. Res..
[14] Yuan Yao,et al. Online Learning as Stochastic Approximation of Regularization Paths: Optimality and Almost-Sure Convergence , 2011, IEEE Transactions on Information Theory.
[15] Elad Hazan,et al. An optimal algorithm for stochastic strongly-convex optimization , 2010, 1006.2425.