暂无分享,去创建一个
Vatsal Sharan | Peter Bailis | Gregory Valiant | Kai Sheng Tai | Vatsal Sharan | G. Valiant | Peter D. Bailis
[1] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[2] Yan Chen,et al. Reversible sketches for efficient and accurate change detection over network data streams , 2004, IMC '04.
[3] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[4] H. Brendan McMahan,et al. Follow-the-Regularized-Leader and Mirror Descent: Equivalence Theorems and L1 Regularization , 2011, AISTATS.
[5] Philippe Flajolet,et al. Approximate counting: A detailed analysis , 1985, BIT.
[6] Mikkel Thorup,et al. Heavy Hitters via Cluster-Preserving Clustering , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).
[7] Ashwin Lall,et al. Streaming Pointwise Mutual Information , 2009, NIPS.
[8] Saurabh Goyal,et al. Resource-efficient Machine Learning in 2 KB RAM for the Internet of Things , 2017, ICML.
[9] Dawn Xiaodong Song,et al. New Streaming Algorithms for Fast Detection of Superspreaders , 2005, NDSS.
[10] Jimmy J. Lin,et al. Summingbird: A Framework for Integrating Batch and Online MapReduce Computations , 2014, Proc. VLDB Endow..
[11] Christopher Ré,et al. Materialization optimizations for feature selection workloads , 2014, SIGMOD Conference.
[12] Kilian Q. Weinberger,et al. Feature hashing for large scale multitask learning , 2009, ICML '09.
[13] Kevin Duh,et al. Streaming Word Embeddings with the Space-Saving Algorithm , 2017, ArXiv.
[14] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[15] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[16] Richard M. Karp,et al. A simple algorithm for finding frequent elements in streams and bags , 2003, TODS.
[17] Prabhakar Raghavan,et al. Randomized rounding: A technique for provably good algorithms and algorithmic proofs , 1985, Comb..
[18] Samuel Madden,et al. MacroBase: Prioritizing Attention in Fast Data , 2016, SIGMOD Conference.
[19] Yiming Yang,et al. RCV1: A New Benchmark Collection for Text Categorization Research , 2004, J. Mach. Learn. Res..
[20] Seth Flaxman,et al. EU regulations on algorithmic decision-making and a "right to explanation" , 2016, ArXiv.
[21] Zachary Chase Lipton. The mythos of model interpretability , 2016, ACM Queue.
[22] Graham Cormode,et al. An improved data stream summary: the count-min sketch and its applications , 2004, J. Algorithms.
[23] Avi Feller,et al. Algorithmic Decision Making and the Cost of Fairness , 2017, KDD.
[24] Emmanuel J. Candès,et al. Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.
[25] Graham Cormode,et al. What's new: finding significant differences in network data streams , 2004, INFOCOM 2004.
[26] Katsiaryna Mirylenka,et al. Conditional heavy hitters: detecting interesting correlations in data streams , 2015, The VLDB Journal.
[27] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[28] Divyakant Agrawal,et al. Medians and beyond: new aggregation techniques for sensor networks , 2004, SenSys '04.
[29] Martin Zinkevich,et al. Online Convex Programming and Generalized Infinitesimal Gradient Ascent , 2003, ICML.
[30] Dimitris Achlioptas,et al. Database-friendly random projections: Johnson-Lindenstrauss with binary coins , 2003, J. Comput. Syst. Sci..
[31] Lu Wang,et al. Quantiles over data streams: experimental comparisons, new analyses, and further improvements , 2016, The VLDB Journal.
[32] Kai-Min Chung,et al. Why simple hash functions work: exploiting the entropy in a data stream , 2008, SODA '08.
[33] Omer Levy,et al. Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.
[34] W. B. Johnson,et al. Extensions of Lipschitz mappings into Hilbert space , 1984 .
[35] Gustavo Alonso,et al. Augmented Sketch: Faster and More Accurate Stream Processing , 2016, SIGMOD Conference.
[36] Divyakant Agrawal,et al. Efficient Computation of Frequent and Top-k Elements in Data Streams , 2005, ICDT.
[37] Larry Carter,et al. Universal Classes of Hash Functions , 1979, J. Comput. Syst. Sci..
[38] Koby Crammer,et al. Online Classification on a Budget , 2003, NIPS.
[39] Dan Suciu,et al. Causality and Explanations in Databases , 2014, Proc. VLDB Endow..
[40] Yoram Singer,et al. Efficient Online and Batch Learning Using Forward Backward Splitting , 2009, J. Mach. Learn. Res..
[41] Rong Jin,et al. Online feature selection for mining big data , 2012, BigMine '12.
[42] Divyakant Agrawal,et al. Fast data stream algorithms using associative memories , 2007, SIGMOD '07.
[43] Rong Jin,et al. Theory of Dual-sparse Regularized Randomized Reduction , 2015, ICML.
[44] David L Donoho,et al. Compressed sensing , 2006, IEEE Transactions on Information Theory.
[45] Martin Wattenberg,et al. Ad click prediction: a view from the trenches , 2013, KDD.
[46] John Langford,et al. Beating the hold-out: bounds for K-fold and progressive cross-validation , 1999, COLT '99.
[47] Sanjeev Khanna,et al. Space-efficient online computation of quantile summaries , 2001, SIGMOD '01.
[48] Mikkel Thorup,et al. The power of simple tabulation hashing , 2010, STOC.
[49] Paul G. Spirakis,et al. Weighted random sampling with a reservoir , 2006, Inf. Process. Lett..
[50] Daniel M. Kane,et al. Sparser Johnson-Lindenstrauss Transforms , 2010, JACM.
[51] D. Sculley,et al. Large-Scale Learning with Less RAM via Randomization , 2013, ICML.
[52] Inderjit S. Dhillon,et al. Fast Prediction for Large-Scale Kernel Machines , 2014, NIPS.
[53] John Langford,et al. Sparse Online Learning via Truncated Gradient , 2008, NIPS.
[54] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[55] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[56] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[57] Rong Jin,et al. Random Projections for Classification: A Recovery Approach , 2014, IEEE Transactions on Information Theory.
[58] Kavé Salamatian,et al. Anomaly extraction in backbone networks using association rules , 2012, TNET.
[59] Shou-De Lin,et al. Feature Engineering and Classifier Ensemble for KDD Cup 2010 , 2010, KDD 2010.
[60] Ohad Shamir,et al. Without-Replacement Sampling for Stochastic Gradient Methods , 2016, NIPS.
[61] Erik D. Demaine,et al. Frequency Estimation of Internet Packet Streams with Limited Space , 2002, ESA.
[62] Moses Charikar,et al. Finding frequent items in data streams , 2002, Theor. Comput. Sci..
[63] Lawrence K. Saul,et al. Identifying suspicious URLs: an application of large-scale online learning , 2009, ICML '09.
[64] Nobuhiro Kaji,et al. Incremental Skip-gram Model with Negative Sampling , 2017, EMNLP.
[65] John C. Duchi,et al. Minimax rates for memory-bounded sparse linear regression , 2015, COLT.
[66] Minlan Yu,et al. Software Defined Traffic Measurement with OpenSketch , 2013, NSDI.
[67] Kilian Q. Weinberger,et al. The Greedy Miser: Learning under Test-time Budgets , 2012, ICML.
[68] Lin Xiao,et al. Dual Averaging Methods for Regularized Stochastic Learning and Online Optimization , 2009, J. Mach. Learn. Res..
[69] Matt J. Kusner,et al. Cost-Sensitive Tree of Classifiers , 2012, ICML.
[70] Samuel Madden,et al. Scorpion: Explaining Away Outliers in Aggregate Queries , 2013, Proc. VLDB Endow..
[71] Patrick Pantel,et al. From Frequency to Meaning: Vector Space Models of Semantics , 2010, J. Artif. Intell. Res..
[72] Prateek Jain,et al. ProtoNN: Compressed and Accurate kNN for Resource-scarce Devices , 2017, ICML.
[73] Yoram Singer,et al. Pegasos: primal estimated sub-gradient solver for SVM , 2011, Math. Program..