Conformal prediction in manifold learning

The paper presents a geometrically motivated view on conformal prediction applied to nonlinear multi-output regression tasks for obtaining valid measure of accuracy of Manifold Learning Regression algorithms. A considered regression task is to estimate an unknown smooth mapping f from q-dimensional inputs x ∈ X to m-dimensional outputs y = f(x) based on training dataset Z(n) consisting of “input-output” pairs {Zi = (xi,yi = f(xi)) , i = 1, 2, . . . , n}. Manifold Learning Regression (MLR) algorithm solves this task using Manifold learning technique. At first, unknown q-dimensional Regression manifold M(f) = {(x, f(x)) ∈ R : x ∈ X ⊂ R}, embedded in ambient (q + m)-dimensional space, is estimated from the training data Z(n), sampled from this manifold. The constructed estimator MMLR, which is also q-dimensional manifold embedded in ambient space R, is close to M in terms of Hausdorff distance. After that, an estimator fMLR of the unknown function f , mapping arbitrary input x ∈ X to output fMLR(x), is constructed as the solution to the equation M(fMLR) = MMLR. Conformal prediction allows constructing a prediction region for an unknown output y = f(x) at Out-of-Sample input point x for a given confidence level using given nonconformity measure, characterizing to which extent an example Z = (x,y) is different from examples in the known dataset Z(n). The paper proposes a new nonconformity measure based on MLR estimators using an analog of Bregman distance.

[1]  L. Bregman The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming , 1967 .

[2]  Alexander P. Kuleshov,et al.  Tangent Bundle Manifold Learning via Grassmann&Stiefel Eigenmaps , 2012, ArXiv.

[3]  Xin Guo,et al.  On the optimality of conditional expectation as a Bregman predictor , 2005, IEEE Trans. Inf. Theory.

[4]  Pascal Frossard,et al.  Tangent space estimation for smooth embeddings of Riemannian manifolds , 2012 .

[5]  Inderjit S. Dhillon,et al.  Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..

[6]  Guohua Pan,et al.  Local Regression and Likelihood , 1999, Technometrics.

[7]  Evgeny Burnaev,et al.  Conformalized Kernel Ridge Regression , 2016, 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA).

[8]  Dong Yu,et al.  Deep Learning: Methods and Applications , 2014, Found. Trends Signal Process..

[9]  Daniel N. Kaslovsky,et al.  Non-Asymptotic Analysis of Tangent Space Perturbation , 2011 .

[10]  Christopher K. I. Williams,et al.  Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning) , 2005 .

[11]  Shun-ichi Amari,et al.  Methods of information geometry , 2000 .

[12]  Evgeny Burnaev,et al.  Adaptive Design of Experiments Based on Gaussian Processes , 2015, SLDS.

[13]  Petra Perner,et al.  Machine Learning and Data Mining in Pattern Recognition , 2009, Lecture Notes in Computer Science.

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  A. Bernstein Data-based Manifold Reconstruction via Tangent Bundle Manifold Learning , 2014 .

[16]  Lawrence Cayton,et al.  Algorithms for manifold learning , 2005 .

[17]  Larry A. Wasserman,et al.  Minimax Manifold Estimation , 2010, J. Mach. Learn. Res..

[18]  Maxim Panov,et al.  Regression on the basis of nonstationary Gaussian processes with Bayesian regularization , 2016 .

[19]  John M. Lee Manifolds and Differential Geometry , 2009 .

[20]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[21]  Yunqian Ma,et al.  Manifold Learning Theory and Applications , 2011 .

[22]  Y. Yanovich Asymptotic Properties of Local Sampling on Manifold , 2016 .

[23]  Harris Papadopoulos,et al.  Regression Conformal Prediction with Nearest Neighbours , 2014, J. Artif. Intell. Res..

[24]  J. Jost Riemannian geometry and geometric analysis , 1995 .

[25]  Evgeny Burnaev,et al.  Gaussian Process Regression for Structured Data Sets , 2015, SLDS.

[26]  Vladimir Vovk,et al.  A tutorial on conformal prediction , 2007, J. Mach. Learn. Res..

[27]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[28]  Vladimir Spokoiny,et al.  Теорема Бернштейна - фон Мизеса для регрессии на основе гауссовских процессов@@@The Bernstein - von Mises theorem for regression based on Gaussian Processes , 2013 .

[29]  Shun-ichi Amari,et al.  Natural Gradient Works Efficiently in Learning , 1998, Neural Computation.

[30]  Evgeny Burnaev,et al.  GTApprox: Surrogate modeling for industrial design , 2016, Adv. Eng. Softw..

[31]  W. Gasarch,et al.  The Book Review Column 1 Coverage Untyped Systems Simple Types Recursive Types Higher-order Systems General Impression 3 Organization, and Contents of the Book , 2022 .

[32]  Evgeny V. Burnaev,et al.  Properties of the posterior distribution of a regression model based on Gaussian random fields , 2013, Autom. Remote. Control..

[33]  Alexander P. Kuleshov,et al.  Manifold Learning in Data Mining Tasks , 2014, MLDM.

[34]  Trevor Hastie,et al.  An Introduction to Statistical Learning , 2013, Springer Texts in Statistics.

[35]  Alexander P. Kuleshov,et al.  Nonlinear multi-output regression on unknown input manifold , 2017, Annals of Mathematics and Artificial Intelligence.

[36]  A. Singer,et al.  Vector diffusion maps and the connection Laplacian , 2011, Communications on pure and applied mathematics.

[37]  X. Huo,et al.  A Survey of Manifold-Based Learning Methods , 2007 .

[38]  Alexander P. Kuleshov,et al.  Extended Regression on Manifolds Estimation , 2016, COPA.