暂无分享,去创建一个
Lawrence Carin | Chenyang Tao | Junya Chen | Zidi Xiu | Benjamin Goldstein | Ricardo Henao | L. Carin | Chenyang Tao | R. Henao | Zidi Xiu | Junya Chen | Benjamin A. Goldstein | B. Goldstein
[1] Saman Ghili,et al. Tiny ImageNet Visual Recognition Challenge , 2014 .
[2] Shakir Mohamed,et al. Variational Inference with Normalizing Flows , 2015, ICML.
[3] Pietro Perona,et al. One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[4] Yang Song,et al. Class-Balanced Loss Based on Effective Number of Samples , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Martial Hebert,et al. Learning Compositional Representations for Few-Shot Recognition , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Aapo Hyvärinen,et al. Nonlinear ICA Using Auxiliary Variables and Generalized Contrastive Learning , 2018, AISTATS.
[7] Don R. Hush,et al. Network constraints and multi-objective optimization for one-class classification , 1996, Neural Networks.
[8] Nitesh V. Chawla,et al. SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..
[9] Dirk P. Kroese,et al. An Efficient Algorithm for Rare-event Probability Estimation, Combinatorial Optimization, and Counting , 2008 .
[10] Jonas Peters,et al. Causal inference by using invariant prediction: identification and confidence intervals , 2015, 1501.01332.
[11] N. Meinshausen,et al. Anchor regression: Heterogeneous data meet causality , 2018, Journal of the Royal Statistical Society: Series B (Statistical Methodology).
[12] Bernhard Schölkopf,et al. On causal and anticausal learning , 2012, ICML.
[13] Iain Murray,et al. Masked Autoregressive Flow for Density Estimation , 2017, NIPS.
[14] Colin Wei,et al. Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss , 2019, NeurIPS.
[15] James T. Kwok,et al. Generalizing from a Few Examples , 2019, ACM Comput. Surv..
[16] Ross B. Girshick,et al. Focal Loss for Dense Object Detection , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[17] Quoc V. Le,et al. AutoAugment: Learning Augmentation Strategies From Data , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Ker-Chau Li,et al. Sliced Inverse Regression for Dimension Reduction , 1991 .
[19] Ya Le,et al. Tiny ImageNet Visual Recognition Challenge , 2015 .
[20] Robert C. Holte,et al. C4.5, Class Imbalance, and Cost Sensitivity: Why Under-Sampling beats Over-Sampling , 2003 .
[21] Vladimir N. Vapnik,et al. The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.
[22] David Lopez-Paz,et al. Invariant Risk Minimization , 2019, ArXiv.
[23] Aapo Hyvärinen,et al. Variational Autoencoders and Nonlinear ICA: A Unifying Framework , 2019, AISTATS.
[24] Aapo Hyvärinen,et al. Nonlinear independent component analysis: Existence and uniqueness results , 1999, Neural Networks.
[25] P. Bühlmann,et al. Invariance, Causality and Robustness , 2018, Statistical Science.
[26] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[27] Amos J. Storkey,et al. Data Augmentation Generative Adversarial Networks , 2017, ICLR 2018.
[28] Bernhard Schölkopf,et al. Invariant Models for Causal Transfer Learning , 2015, J. Mach. Learn. Res..
[29] Pierre Comon. Independent component analysis - a new concept? signal processing , 1994 .
[30] Ross B. Girshick,et al. Reducing Overfitting in Deep Networks by Decorrelating Representations , 2015, ICLR.
[31] Feng Guo,et al. Driver crash risk factors and prevalence evaluation using naturalistic driving data , 2016, Proceedings of the National Academy of Sciences.
[32] Masashi Sugiyama,et al. Few-shot Domain Adaptation by Causal Mechanism Transfer , 2020, ICML.
[33] Frank D. Wood,et al. Learning Disentangled Representations with Semi-Supervised Deep Generative Models , 2017, NIPS.
[34] Haibo He,et al. ADASYN: Adaptive synthetic sampling approach for imbalanced learning , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).
[35] António M Lopes,et al. Rare and extreme events: the case of COVID-19 pandemic , 2020, Nonlinear dynamics.
[36] Stella X. Yu,et al. Large-Scale Long-Tailed Recognition in an Open World , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[37] Razvan Pascanu,et al. Natural Neural Networks , 2015, NIPS.
[38] Constantine Bekas,et al. BAGAN: Data Augmentation with Balancing GAN , 2018, ArXiv.
[39] Dacheng Tao,et al. Orthogonal Deep Neural Networks , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[40] A. Choromańska. Extreme Multi Class Classification , 2013 .
[41] Pierre Comon,et al. Independent component analysis, A new concept? , 1994, Signal Process..
[42] Yang Song,et al. The iNaturalist Species Classification and Detection Dataset , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[43] David Duvenaud,et al. Neural Ordinary Differential Equations , 2018, NeurIPS.
[44] Charles Elkan,et al. The Foundations of Cost-Sensitive Learning , 2001, IJCAI.
[45] Suvrit Sra,et al. Strength from Weakness: Fast Learning Using Weak Supervision , 2020, ICML.
[46] Taghi M. Khoshgoftaar,et al. A survey on Image Data Augmentation for Deep Learning , 2019, Journal of Big Data.
[47] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[48] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[49] Xiaohan Chen,et al. Can We Gain More from Orthogonality Regularizations in Training Deep CNNs? , 2018, NeurIPS.
[50] K. Strimmer,et al. Optimal Whitening and Decorrelation , 2015, 1512.00809.
[51] Erkki Oja,et al. Independent component analysis: algorithms and applications , 2000, Neural Networks.
[52] Lei Huang,et al. Decorrelated Batch Normalization , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[53] Alexandra Chouldechova,et al. The Frontiers of Fairness in Machine Learning , 2018, ArXiv.
[54] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[55] Sankha Subhra Mullick,et al. Generative Adversarial Minority Oversampling , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[56] Gary King,et al. Logistic Regression in Rare Events Data , 2001, Political Analysis.
[57] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[58] Cesare Alippi,et al. Credit Card Fraud Detection: A Realistic Modeling and a Novel Learning Strategy , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[59] Christina Heinze-Deml,et al. Invariant Causal Prediction for Nonlinear Models , 2017, Journal of Causal Inference.
[60] Alexander Binder,et al. Deep One-Class Classification , 2018, ICML.
[61] Zhangyang Wang,et al. Can We Gain More from Orthogonality Regularizations in Training Deep Networks? , 2018, NeurIPS.
[62] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[63] Martial Hebert,et al. Learning to Model the Tail , 2017, NIPS.
[64] Dimitris Kanellopoulos,et al. Handling imbalanced datasets: A review , 2006 .
[65] Zachary C. Lipton,et al. What is the Effect of Importance Weighting in Deep Learning? , 2018, ICML.
[66] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).