Calibrated Surrogate Losses for Adversarially Robust Classification
暂无分享,去创建一个
[1] Csaba Szepesvári,et al. Cost-sensitive Multiclass Classification Risk Bounds , 2013, ICML.
[2] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.
[3] Pradeep Ravikumar,et al. On NDCG Consistency of Listwise Ranking Methods , 2011, AISTATS.
[4] David Tse,et al. A Minimax Approach to Supervised Learning , 2016, NIPS.
[5] Philip M. Long,et al. Consistency versus Realizable H-Consistency for Multiclass Classification , 2013, ICML.
[6] Po-Ling Loh,et al. Adversarial Risk Bounds via Function Transformation , 2018 .
[7] Didier Aussel,et al. Subdifferential characterization of quasiconvexity and convexity , 1994 .
[8] Mathieu Blondel,et al. Structured Prediction with Projection Oracles , 2019, NeurIPS.
[9] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[10] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[11] Matthias Hein,et al. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation , 2017, NIPS.
[12] Shie Mannor,et al. Robustness and generalization , 2010, Machine Learning.
[13] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[14] Gang Niu,et al. Does Distributionally Robust Supervised Learning Give Robust Classifiers? , 2016, ICML.
[15] Prasad Raghavendra,et al. Agnostic Learning of Monomials by Halfspaces Is Hard , 2009, 2009 50th Annual IEEE Symposium on Foundations of Computer Science.
[16] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[17] Masashi Sugiyama,et al. On Symmetric Losses for Learning from Corrupted Labels , 2019, ICML.
[18] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[19] Ingo Steinwart. How to Compare Different Loss Functions and Their Risks , 2007 .
[20] Shivani Agarwal,et al. Convex Calibration Dimension for Multiclass Loss Matrices , 2014, J. Mach. Learn. Res..
[21] Richard Nock,et al. Monge beats Bayes: Hardness Results for Adversarial Training , 2018, ICML.
[22] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[23] John C. Duchi,et al. Stochastic Gradient Methods for Distributionally Robust Optimization with f-divergences , 2016, NIPS.
[24] Jason Weston,et al. Trading convexity for scalability , 2006, ICML.
[25] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[26] Yishay Mansour,et al. Domain Adaptation: Learning Bounds and Algorithms , 2009, COLT.
[27] Greg Yang,et al. Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers , 2019, NeurIPS.
[28] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[29] Brian D. Ziebart,et al. Adversarial Multiclass Classification: A Risk Minimization Perspective , 2016, NIPS.
[30] Felix Schlenk,et al. Proof of Theorem 3 , 2005 .
[31] Frank Nielsen,et al. Loss factorization, weakly supervised learning and label noise robustness , 2016, ICML.
[32] Masashi Sugiyama,et al. Calibrated Surrogate Maximization of Linear-fractional Utility in Binary Classification , 2019, AISTATS.
[33] Francis R. Bach,et al. On Structured Prediction Theory with Calibrated Convex Surrogate Losses , 2017, NIPS.
[34] John C. Duchi,et al. Variance-based Regularization with Convex Objectives , 2016, NIPS.
[35] Michael I. Jordan,et al. Convexity, Classification, and Risk Bounds , 2006 .
[36] Michael I. Jordan,et al. On the Consistency of Ranking Algorithms , 2010, ICML.
[37] F. Clarke. Optimization And Nonsmooth Analysis , 1983 .
[38] Rocco A. Servedio,et al. Random classification noise defeats all convex potential boosters , 2008, ICML '08.
[39] Tong Zhang. Statistical behavior and consistency of classification methods based on convex risk minimization , 2003 .
[40] Tamir Hazan,et al. Direct Loss Minimization for Structured Prediction , 2010, NIPS.
[41] Matthew J. Holland. Classification using margin pursuit , 2018, AISTATS.
[42] Constantine Caramanis,et al. Theory and Applications of Robust Optimization , 2010, SIAM Rev..
[43] Zhi-Hua Zhou,et al. On the Consistency of Multi-Label Learning , 2011, COLT.
[44] Alexander J. Smola,et al. Second Order Cone Programming Approaches for Handling Missing and Uncertain Data , 2006, J. Mach. Learn. Res..
[45] Yi Lin. A note on margin-based loss functions in classification , 2004 .
[46] Mark D. Reid,et al. Composite Binary Losses , 2009, J. Mach. Learn. Res..
[47] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[48] Ambuj Tewari,et al. Convex Calibrated Surrogates for Low-Rank Loss Matrices with Applications to Subset Ranking Losses , 2013, NIPS.
[49] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.
[50] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[51] Zhi-Hua Zhou,et al. On the Consistency of AUC Pairwise Optimization , 2012, IJCAI.
[52] Shivani Agarwal,et al. Classification Calibration Dimension for General Multiclass Losses , 2012, NIPS.
[53] Uri Shaham,et al. Understanding adversarial training: Increasing local stability of supervised models through robust optimization , 2015, Neurocomputing.
[54] Aritra Ghosh,et al. Making risk minimization tolerant to label noise , 2014, Neurocomputing.
[55] Daniel M. Kane,et al. Nearly Tight Bounds for Robust Proper Learning of Halfspaces with a Margin , 2019, NeurIPS.
[56] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[57] Jean-Philippe Vial,et al. Robust Optimization , 2021, ICORES.
[58] Nuno Vasconcelos,et al. On the Design of Loss Functions for Classification: theory, robustness to outliers, and SavageBoost , 2008, NIPS.
[59] D. Angluin,et al. Learning From Noisy Examples , 1988, Machine Learning.
[60] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[61] Michael I. Jordan,et al. A Robust Minimax Approach to Classification , 2003, J. Mach. Learn. Res..
[62] Ambuj Tewari,et al. On the Consistency of Multiclass Classification Methods , 2007, J. Mach. Learn. Res..
[63] Csaba Szepesvári,et al. Multiclass Classification Calibration Functions , 2016, ArXiv.
[64] François Laviolette,et al. A PAC-Bayesian Approach for Domain Adaptation with Specialization to Linear Classifiers , 2013, ICML.
[65] E. Hüllermeier,et al. Consistent multilabel ranking through univariate loss minimization , 2012, ICML 2012.
[66] Koby Crammer,et al. A theory of learning from different domains , 2010, Machine Learning.
[67] Yuchen Zhang,et al. Bridging Theory and Algorithm for Domain Adaptation , 2019, ICML.
[68] C. Scott. Calibrated asymmetric surrogate losses , 2012 .
[69] Masashi Sugiyama,et al. Unsupervised Domain Adaptation Based on Source-guided Discrepancy , 2018, AAAI.
[70] Hisashi Kashima,et al. Theoretical evidence for adversarial robustness through randomization: the case of the Exponential family , 2019, NeurIPS.