暂无分享,去创建一个
[1] Joseph M. Hellerstein,et al. GraphLab: A New Framework For Parallel Machine Learning , 2010, UAI.
[2] Cristina V. Lopes. Map Reduce , 2020, Exercises in Programming Style.
[3] John C. Duchi,et al. Distributed delayed stochastic optimization , 2011, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[4] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[5] H. E. Solberg,et al. Detection of outliers in reference distributions: performance of Horn's algorithm. , 2005, Clinical chemistry.
[6] Fred Roosta,et al. DINGO: Distributed Newton-Type Method for Gradient-Norm Optimization , 2019, NeurIPS.
[7] Sébastien Bubeck,et al. Convex Optimization: Algorithms and Complexity , 2014, Found. Trends Mach. Learn..
[8] Nenghai Yu,et al. Asynchronous Stochastic Gradient Descent with Delay Compensation , 2016, ICML.
[9] Xin Yao,et al. Diversity creation methods: a survey and categorisation , 2004, Inf. Fusion.
[10] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[11] Emile Fiesler,et al. Neural Networks with Adaptive Learning Rate and Momentum Terms , 1995 .
[12] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[13] Anuraganand Sharma,et al. Guided Stochastic Gradient Descent Algorithm for inconsistent datasets , 2018, Appl. Soft Comput..
[14] Yuchen Zhang,et al. DiSCO: Distributed Optimization for Self-Concordant Empirical Loss , 2015, ICML.
[15] Kunle Olukotun,et al. Map-Reduce for Machine Learning on Multicore , 2006, NIPS.
[16] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[17] Bin-Da Liu,et al. A backpropagation algorithm with adaptive learning rate and momentum coefficient , 2002, Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290).
[18] Shusen Wang,et al. GIANT: Globally Improved Approximate Newton Method for Distributed Optimization , 2017, NeurIPS.
[19] Martti Juhola,et al. Informal identification of outliers in medical data , 2000 .
[20] A. Asuncion,et al. UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .
[21] Dr. Zbigniew Michalewicz,et al. How to Solve It: Modern Heuristics , 2004 .
[22] Alexander J. Smola,et al. Parallelized Stochastic Gradient Descent , 2010, NIPS.
[23] Heekuck Oh,et al. Neural Networks for Pattern Recognition , 1993, Adv. Comput..
[24] Martin Jaggi,et al. Sparsified SGD with Memory , 2018, NeurIPS.
[25] Mark W. Schmidt,et al. A simpler approach to obtaining an O(1/t) convergence rate for the projected stochastic subgradient method , 2012, ArXiv.
[26] Qihui Wu,et al. A survey of machine learning for big data processing , 2016, EURASIP Journal on Advances in Signal Processing.
[27] Ke Tang,et al. Stochastic Gradient Descent for Nonconvex Learning Without Bounded Gradient Assumptions , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[28] Dan Alistarh,et al. The Convergence of Stochastic Gradient Descent in Asynchronous Shared Memory , 2018, PODC.
[29] M. Marques,et al. Recent advances and applications of machine learning in solid-state materials science , 2019, npj Computational Materials.
[30] Dimitrios I. Fotiadis,et al. Machine learning applications in cancer prognosis and prediction , 2014, Computational and structural biotechnology journal.
[31] Raghu Meka. CS289ML: Notes on convergence of gradient descent , 2016 .