暂无分享,去创建一个
[1] Kun Li,et al. The MADlib Analytics Library or MAD Skills, the SQL , 2012, Proc. VLDB Endow..
[2] Shirish Tatikonda,et al. SystemML: Declarative machine learning on MapReduce , 2011, 2011 IEEE 27th International Conference on Data Engineering.
[3] Eric P. Xing,et al. GeePS: scalable deep learning on distributed GPUs with a GPU-specialized parameter server , 2016, EuroSys.
[4] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[5] Christopher Ré,et al. DimmWitted: A Study of Main-Memory Statistical Analytics , 2014, Proc. VLDB Endow..
[6] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition , 2012 .
[7] Diego Klabjan,et al. Classification-Based Financial Markets Prediction Using Deep Neural Networks , 2016, Algorithmic Finance.
[8] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Dan Olteanu,et al. Learning Linear Regression Models over Factorized Joins , 2016, SIGMOD Conference.
[10] Florin Rusu,et al. Scalable Asynchronous Gradient Descent Optimization for Out-of-Core Models , 2017, Proc. VLDB Endow..
[11] Gang Chen,et al. Database Meets Deep Learning: Challenges and Opportunities , 2016, SGMD.
[12] Samy Bengio,et al. Revisiting Distributed Synchronous SGD , 2016, ArXiv.
[13] Christopher Ré,et al. Towards a unified architecture for in-RDBMS analytics , 2012, SIGMOD Conference.
[14] Michael I. Jordan,et al. Estimation, Optimization, and Parallelism when Data is Sparse , 2013, NIPS.
[15] Joel A. Tropp,et al. Factoring nonnegative matrices with linear programs , 2012, NIPS.
[16] Sanjay Chawla,et al. A Cost-based Optimizer for Gradient Descent Optimization , 2017, SIGMOD Conference.
[17] Jeffrey F. Naughton,et al. To Join or Not to Join?: Thinking Twice about Joins before Feature Selection , 2016, SIGMOD Conference.
[18] Kunle Olukotun,et al. Understanding and optimizing asynchronous low-precision stochastic gradient descent , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).
[19] Florin Rusu,et al. Speculative Approximations for Terascale Distributed Gradient Descent Optimization , 2015, DanaC@SIGMOD.
[20] Florin Rusu,et al. Scalable I/O-bound parallel incremental gradient descent for big data analytics in GLADE , 2013, DanaC '13.
[21] Ioannis Mitliagkas,et al. Omnivore: An Optimizer for Multi-device Deep Learning on CPUs and GPUs , 2016, ArXiv.
[22] Nenghai Yu,et al. Asynchronous Stochastic Gradient Descent with Delay Compensation , 2016, ICML.
[23] Keshav Pingali,et al. Stochastic gradient descent on GPUs , 2015, GPGPU@PPoPP.
[24] John Langford,et al. A reliable effective terascale linear learning system , 2011, J. Mach. Learn. Res..
[25] Alexander J. Smola,et al. Parallelized Stochastic Gradient Descent , 2010, NIPS.
[26] Tsuyoshi Murata,et al. {m , 1934, ACML.
[27] Christopher R'e,et al. Caffe con Troll: Shallow Ideas to Speed Up Deep Learning , 2015, DanaC@SIGMOD.
[28] Dimitri P. Bertsekas,et al. Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey , 2015, ArXiv.
[29] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[30] Christopher Ré,et al. High Performance Parallel Stochastic Gradient Descent in Shared Memory , 2016, 2016 IEEE International Parallel and Distributed Processing Symposium (IPDPS).
[31] Tim Kraska,et al. MLI: An API for Distributed Machine Learning , 2013, 2013 IEEE 13th International Conference on Data Mining.
[32] Grgoire Montavon,et al. Neural Networks: Tricks of the Trade , 2012, Lecture Notes in Computer Science.
[33] Florin Rusu,et al. Dot-Product Join: Scalable In-Database Linear Algebra for Big Model Analytics , 2017, SSDBM.
[34] Haris Pozidis,et al. Large-Scale Stochastic Learning Using GPUs , 2017, 2017 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW).
[35] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[36] Stephen J. Wright,et al. An asynchronous parallel stochastic coordinate descent algorithm , 2013, J. Mach. Learn. Res..
[37] Yun Liang,et al. CuMF_SGD: Parallelized Stochastic Gradient Descent for Matrix Factorization on GPUs , 2017, HPDC.
[38] Samuel Williams,et al. Optimization of sparse matrix-vector multiplication on emerging multicore platforms , 2009, Parallel Comput..
[39] Trishul M. Chilimbi,et al. Project Adam: Building an Efficient and Scalable Deep Learning Training System , 2014, OSDI.
[40] Pradeep Dubey,et al. Debunking the 100X GPU vs. CPU myth: an evaluation of throughput computing on CPU and GPU , 2010, ISCA.
[41] Jeffrey F. Naughton,et al. Learning Generalized Linear Models Over Normalized Data , 2015, SIGMOD Conference.