Blessing of dimensionality at the edge and geometry of few-shot learning

Abstract In this paper we present theory and algorithms enabling classes of Artificial Intelligence (AI) systems to continuously and incrementally improve with a-priori quantifiable guarantees – or more specifically remove classification errors – over time. This is distinct from state-of-the-art machine learning, AI, and software approaches. The theory enables building few-shot AI correction algorithms and provides conditions justifying their successful application. Another feature of this approach is that, in the supervised setting, the computational complexity of training is linear in the number of training samples. At the time of classification, the computational complexity is bounded by few inner product calculations. Moreover, the implementation is shown to be very scalable. This makes it viable for deployment in applications where computational power and memory are limited, such as embedded environments. It enables the possibility for fast on-line optimisation using improved training samples. The approach is based on the concentration of measure effects and stochastic separation theorems and is illustrated with an example on the identification faulty processes in Computer Numerical Control (CNC) milling and with a case study on adaptive removal of false positives in an industrial video surveillance and analytics system.

[1]  Rauf Izmailov,et al.  Knowledge transfer in SVM and neural networks , 2017, Annals of Mathematics and Artificial Intelligence.

[2]  Bogdan Grechuk,et al.  General stochastic separation theorems with optimal bounds , 2020, ArXiv.

[3]  Lionel M. Ni,et al.  Generalizing from a Few Examples , 2020, ACM Comput. Surv..

[4]  Yang Song,et al.  Improving the Robustness of Deep Neural Networks via Stability Training , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Andrei Zinovyev,et al.  Independent Component Analysis for Unraveling the Complexity of Cancer Omics Datasets , 2019, International journal of molecular sciences.

[6]  Ivan Tyukin,et al.  Correction of AI systems by linear discriminants: Probabilistic foundations , 2018, Inf. Sci..

[7]  Ivan Tyukin,et al.  The Blessing of Dimensionality: Separation Theorems in the Thermodynamic Limit , 2016, ArXiv.

[8]  Dianhui Wang,et al.  Stochastic Configuration Networks: Fundamentals and Algorithms , 2017, IEEE Transactions on Cybernetics.

[9]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[10]  Ivan Tyukin,et al.  Blessing of dimensionality: mathematical foundations of the statistical physics of data , 2018, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[11]  Lorien Y. Pratt,et al.  Discriminability-Based Transfer between Neural Networks , 1992, NIPS.

[12]  Eliza Strickland,et al.  IBM Watson, heal thyself: How IBM overpromised and underdelivered on AI health care , 2019, IEEE Spectrum.

[13]  Richard S. Zemel,et al.  Prototypical Networks for Few-shot Learning , 2017, NIPS.

[14]  Bodo Rosenhahn,et al.  Expanding object detector's Horizon: Incremental learning framework for object detection in videos , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  A. Giuliani,et al.  System Biology Approach: Gene Network Analysis for Muscular Dystrophy. , 2018, Methods in molecular biology.

[16]  Haiping Lu,et al.  A survey of multilinear subspace learning for tensor data , 2011, Pattern Recognit..

[17]  Konstantin I. Sofeikov,et al.  Knowledge Transfer Between Artificial Intelligence Systems , 2017, Front. Neurorobot..

[18]  Cordelia Schmid,et al.  Learning object class detectors from weakly annotated video , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[19]  Ivan Tyukin,et al.  Fast Construction of Correcting Ensembles for Legacy Artificial Intelligence Systems: Algorithms and a Case Study , 2018, Inf. Sci..

[20]  Ivan Tyukin,et al.  Augmented Artificial Intelligence: a Conceptual Framework , 2018, ArXiv.

[21]  Oriol Vinyals,et al.  Matching Networks for One Shot Learning , 2016, NIPS.

[22]  Lars Kai Hansen,et al.  Neural Network Ensembles , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[23]  Mark Harman,et al.  Machine Learning Testing: Survey, Landscapes and Horizons , 2019, IEEE Transactions on Software Engineering.

[24]  Olivier Chapelle,et al.  Training a Support Vector Machine in the Primal , 2007, Neural Computation.

[25]  György Elekes,et al.  A geometric inequality and the complexity of computing volume , 1986, Discret. Comput. Geom..

[26]  Ivan Tyukin,et al.  One-trial correction of legacy AI systems and stochastic separation theorems , 2019, Inf. Sci..

[27]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[28]  Ivan Tyukin,et al.  High-Dimensional Brain in a High-Dimensional World: Blessing of Dimensionality , 2020, Entropy.

[29]  Liming Chen,et al.  Knowledge Transfer in Vision Recognition , 2020, ACM Comput. Surv..

[30]  Ivan Tyukin,et al.  Stochastic Separation Theorems , 2017, Neural Networks.

[31]  M. Gromov Isoperimetry of waists and concentration of maps , 2003 .

[32]  Andrei Yu. Zinovyev,et al.  Estimating the effective dimension of large biological datasets using Fisher separability analysis , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[33]  Paul C. Kainen,et al.  Utilizing Geometric Anomalies of High Dimension: When Complexity Makes Computation Easier , 1997 .

[34]  Hugo Guterman,et al.  Functional Safety Verification for Autonomous UGVs—Methodology Presentation and Implementation on a Full-Scale System , 2019, IEEE Transactions on Intelligent Vehicles.

[35]  Xin Sun,et al.  Evaluation and accurate diagnoses of pediatric diseases using artificial intelligence , 2019, Nature Medicine.

[36]  Nan Li,et al.  Game Theoretic Modeling of Driver and Vehicle Interactions for Verification and Validation of Autonomous Vehicle Control Systems , 2016, IEEE Transactions on Control Systems Technology.

[37]  Ivan Tyukin,et al.  The unreasonable effectiveness of small neural ensembles in high-dimensional brain , 2018, Physics of life reviews.

[38]  R. Irizarry,et al.  Missing data and technical variability in single‐cell RNA‐sequencing experiments , 2018, Biostatistics.

[39]  Peihua Gu,et al.  Recent development in CNC machining of freeform surfaces: A state-of-the-art review , 2010, Comput. Aided Des..

[40]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[41]  Tamas Haidegger,et al.  Highly Automated Vehicles and Self-Driving Cars [Industry Tutorial] , 2018, IEEE Robotics & Automation Magazine.

[42]  Bradley P. Coe,et al.  Copy number variation detection and genotyping from exome sequence data , 2012, Genome research.

[43]  Peter E. Hart,et al.  The condensed nearest neighbor rule (Corresp.) , 1968, IEEE Trans. Inf. Theory.

[44]  Gaétan Hains,et al.  Towards formal methods and software engineering for deep learning: Security, safety and productivity for dl systems development , 2018, 2018 Annual IEEE International Systems Conference (SysCon).

[45]  Paul C. Kainen,et al.  Quasiorthogonal dimension of euclidean spaces , 1993 .

[46]  Ivan Tyukin,et al.  Approximation with random bases: Pro et Contra , 2015, Inf. Sci..

[47]  Alessandro Laio,et al.  Intrinsic dimension of data representations in deep neural networks , 2019, NeurIPS.