暂无分享,去创建一个
Andreas Krause | Sebastian Curi | Kfir Y. Levy | Zalán Borsos | Andreas Krause | K. Levy | Sebastian Curi | Zalán Borsos
[1] Elad Hazan,et al. Introduction to Online Convex Optimization , 2016, Found. Trends Optim..
[2] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[3] Yoram Singer,et al. Efficient projections onto the l1-ball for learning in high dimensions , 2008, ICML '08.
[4] Andreas Krause,et al. Online Variance Reduction for Stochastic Optimization , 2018, COLT.
[5] Zeyuan Allen Zhu,et al. Even Faster Accelerated Coordinate Descent Using Non-Uniform Sampling , 2015, ICML.
[6] Patrick Thiran,et al. Stochastic Optimization with Bandit Sampling , 2017, ArXiv.
[7] D. Freedman. On Tail Probabilities for Martingales , 1975 .
[8] Martin Zinkevich,et al. Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.
[9] Peter Auer,et al. The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..
[10] Simon Haykin,et al. GradientBased Learning Applied to Document Recognition , 2001 .
[11] Hedvig Kjellstrom,et al. Determinantal Point Processes for Mini-Batch Diversification , 2017, UAI 2017.
[12] Guillaume Bouchard,et al. Online Learning to Sample , 2015, 1506.09016.
[13] Guillaume Bouchard,et al. Accelerating Stochastic Gradient Descent via Online Learning to Sample , 2015, ArXiv.
[14] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[15] JoachimsThorsten,et al. KDD-Cup 2004 , 2004 .
[16] John C. Duchi,et al. Adaptive Sampling Probabilities for Non-Smooth Optimization , 2017, ICML.
[17] D. Sculley,et al. Web-scale k-means clustering , 2010, WWW '10.
[18] Volkan Cevher,et al. Faster Coordinate Descent via Adaptive Importance Sampling , 2017, AISTATS.
[19] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[20] Ambuj Tewari,et al. On the Generalization Ability of Online Strongly Convex Programming Algorithms , 2008, NIPS.
[21] Tong Zhang,et al. Accelerating Minibatch Stochastic Gradient Descent using Stratified Sampling , 2014, ArXiv.
[22] Elad Hazan,et al. Logarithmic regret algorithms for online convex optimization , 2006, Machine Learning.
[23] Tyler B. Johnson,et al. Training Deep Models Faster with Robust, Approximate Importance Sampling , 2018, NeurIPS.
[24] Andreas Krause,et al. The next big one: Detecting earthquakes and other rare events from community-based sensors , 2011, Proceedings of the 10th ACM/IEEE International Conference on Information Processing in Sensor Networks.
[25] Sergei Vassilvitskii,et al. k-means++: the advantages of careful seeding , 2007, SODA '07.
[26] Martin Jaggi,et al. Safe Adaptive Importance Sampling , 2017, NIPS.
[27] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[28] Tom Schaul,et al. Prioritized Experience Replay , 2015, ICLR.
[29] Deanna Needell,et al. Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm , 2013, Mathematical Programming.
[30] Ben Taskar,et al. Determinantal Point Processes for Machine Learning , 2012, Found. Trends Mach. Learn..
[31] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[32] Frank Hutter,et al. Online Batch Selection for Faster Training of Neural Networks , 2015, ArXiv.
[33] Tong Zhang,et al. Stochastic Optimization with Importance Sampling for Regularized Loss Minimization , 2014, ICML.
[34] Patrick Thiran,et al. Coordinate Descent with Bandit Sampling , 2017, NeurIPS.
[35] François Fleuret,et al. Not All Samples Are Created Equal: Deep Learning with Importance Sampling , 2018, ICML.
[36] D. Bertsekas,et al. TWO-METRIC PROJECTION METHODS FOR CONSTRAINED OPTIMIZATION* , 1984 .