PARMAC: DISTRIBUTED OPTIMISATION
暂无分享,去创建一个
[1] G. McLachlan,et al. The EM algorithm and extensions , 1996 .
[2] Anthony Skjellum,et al. Using MPI: portable parallel programming with the message-passing interface, 2nd Edition , 1999, Scientific and engineering computation series.
[3] Christopher K. I. Williams,et al. Using the Nyström Method to Speed Up Kernel Machines , 2000, NIPS.
[4] Stephen J. Wright,et al. Numerical Optimization (Springer Series in Operations Research and Financial Engineering) , 2000 .
[5] David P. Anderson,et al. SETI@home: an experiment in public-resource computing , 2002, CACM.
[6] Nicolas Le Roux,et al. Out-of-Sample Extensions for LLE, Isomap, MDS, Eigenmaps, and Spectral Clustering , 2003, NIPS.
[7] Petros Drineas,et al. On the Nyström Method for Approximating a Gram Matrix for Improved Kernel-Based Learning , 2005, J. Mach. Learn. Res..
[8] Léon Bottou,et al. The Tradeoffs of Large Scale Learning , 2007, NIPS.
[9] Ameet Talwalkar,et al. Large-scale manifold learning , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.
[10] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[11] Scott Shenker,et al. Spark: Cluster Computing with Working Sets , 2010, HotCloud.
[12] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[13] Matthijs Douze,et al. Searching in one billion vectors: Re-rank with source coding , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[14] Joseph K. Bradley,et al. Parallel Coordinate Descent for L1-Regularized Loss Minimization , 2011, ICML.
[15] Cordelia Schmid,et al. Product Quantization for Nearest Neighbor Search , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[16] Stephen P. Boyd,et al. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..
[17] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[18] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups , 2012, IEEE Signal Processing Magazine.
[19] Carlos Guestrin,et al. Distributed GraphLab : A Framework for Machine Learning and Data Mining in the Cloud , 2012 .
[20] Alexander G. Gray,et al. Stochastic Alternating Direction Method of Multipliers , 2013, ICML.
[21] Marc'Aurelio Ranzato,et al. Building high-level features using large scale unsupervised learning , 2011, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[22] Miguel Á. Carreira-Perpiñán,et al. Locally Linear Landmarks for Large-Scale Manifold Learning , 2013, ECML/PKDD.
[23] Kristen Grauman,et al. Learning Binary Hash Codes for Large-Scale Image Search , 2013, Machine Learning for Computer Vision.
[24] Svetlana Lazebnik,et al. Iterative quantization: A procrustean approach to learning binary codes , 2011, CVPR 2011.
[25] Tao Wang,et al. Deep learning with COTS HPC systems , 2013, ICML.
[26] Dong Yu,et al. 1-bit stochastic gradient descent and its application to data-parallel distributed training of speech DNNs , 2014, INTERSPEECH.
[27] Miguel Á. Carreira-Perpiñán,et al. Distributed optimization of deeply nested systems , 2012, AISTATS.
[28] Volkan Cevher,et al. Convex Optimization for Big Data: Scalable, randomized, and parallel algorithms for big data analytics , 2014, IEEE Signal Processing Magazine.
[29] James T. Kwok,et al. Asynchronous Distributed ADMM for Consensus Optimization , 2014, ICML.
[30] Miguel Á. Carreira-Perpiñán,et al. Joint optimization of mapping and classifier using auxiliary coordinates , 2014 .
[31] Miguel Á. Carreira-Perpiñán,et al. Hashing with binary autoencoders , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Stephen J. Wright. Coordinate descent algorithms , 2015, Mathematical Programming.
[33] Stephen J. Wright,et al. Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties , 2014, SIAM J. Optim..
[34] Yaoliang Yu,et al. Petuum: A New Platform for Distributed Machine Learning on Big Data , 2015, IEEE Trans. Big Data.
[35] Dimitri P. Bertsekas,et al. Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey , 2015, ArXiv.
[36] Zheng Xu,et al. Training Neural Networks Without Gradients: A Scalable ADMM Approach , 2016, ICML.
[37] Miguel Á. Carreira-Perpiñán,et al. The Variational Nystrom method for large-scale spectral problems , 2016, ICML.
[38] Peter Richtárik,et al. Distributed Coordinate Descent Method for Learning with Big Data , 2013, J. Mach. Learn. Res..
[39] Alex Graves,et al. Decoupled Neural Interfaces using Synthetic Gradients , 2016, ICML.
[40] Daniel Kifer,et al. Conducting Credit Assignment by Aligning Local Representations , 2018, 1803.01834.
[41] Miguel Á. Carreira-Perpiñán,et al. ParMAC: distributed optimisation of nested functions, with application to learning binary autoencoders , 2016, MLSys.