ManifoldNet: A Deep Network Framework for Manifold-valued Data

Deep neural networks have become the main work horse for many tasks involving learning from data in a variety of applications in Science and Engineering. Traditionally, the input to these networks lie in a vector space and the operations employed within the network are well defined on vector-spaces. In the recent past, due to technological advances in sensing, it has become possible to acquire manifold-valued data sets either directly or indirectly. Examples include but are not limited to data from omnidirectional cameras on automobiles, drones etc., synthetic aperture radar imaging, diffusion magnetic resonance imaging, elastography and conductance imaging in the Medical Imaging domain and others. Thus, there is need to generalize the deep neural networks to cope with input data that reside on curved manifolds where vector space operations are not naturally admissible. In this paper, we present a novel theoretical framework to generalize the widely popular convolutional neural networks (CNNs) to high dimensional manifold-valued data inputs. We call these networks, ManifoldNets. In ManifoldNets, convolution operation on data residing on Riemannian manifolds is achieved via a provably convergent recursive computation of the weighted Frechet Mean (wFM) of the given data, where the weights makeup the convolution mask, to be learned. Further, we prove that the proposed wFM layer achieves a contraction mapping and hence ManifoldNet does not need the non-linear ReLU unit used in standard CNNs. We present experiments, using the ManifoldNet framework, to achieve dimensionality reduction by computing the principal linear subspaces that naturally reside on a Grassmannian. The experimental results demonstrate the efficacy of ManifoldNets in the context of classification and reconstruction accuracy.

[1]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[2]  Xiaowei Zhou,et al.  Polar Transformer Networks , 2017, ICLR.

[3]  Alex Graves,et al.  Conditional Image Generation with PixelCNN Decoders , 2016, NIPS.

[4]  Risi Kondor,et al.  On the Generalization of Equivariance and Convolution in Neural Networks to the Action of Compact Groups , 2018, ICML.

[5]  Karl-Theodor Sturm,et al.  Probability Measures on Metric Spaces of Nonpositive Curvature , 2003 .

[6]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[7]  Sander Dieleman,et al.  Rotation-invariant convolutional neural networks for galaxy morphology prediction , 2015, ArXiv.

[8]  Joan Bruna,et al.  Deep Convolutional Networks on Graph-Structured Data , 2015, ArXiv.

[9]  Rudrasis Chakraborty,et al.  A CNN for homogneous Riemannian manifolds with applications to Neuroimaging , 2018, 1805.05487.

[10]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[11]  Søren Hauberg,et al.  Scalable Robust Principal Component Analysis Using Grassmann Averages , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Prafulla Dhariwal,et al.  Glow: Generative Flow with Invertible 1x1 Convolutions , 2018, NeurIPS.

[14]  Stéphane Mallat,et al.  Invariant Scattering Convolution Networks , 2012, IEEE transactions on pattern analysis and machine intelligence.

[15]  Volker Tresp,et al.  Tensor-Train Recurrent Neural Networks for Video Classification , 2017, ICML.

[16]  Pierre Vandergheynst,et al.  Geodesic Convolutional Neural Networks on Riemannian Manifolds , 2015, 2015 IEEE International Conference on Computer Vision Workshop (ICCVW).

[17]  Pedro M. Domingos,et al.  Deep Symmetry Networks , 2014, NIPS.

[18]  Søren Hauberg,et al.  Intrinsic Grassmann Averages for Online Linear and Robust Subspace Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  David Groisser Newton's method, zeroes of vector fields, and the Riemannian center of mass , 2004, Adv. Appl. Math..

[20]  Rudrasis Chakraborty,et al.  H-CNNs: Convolutional Neural Networks for Riemannian Homogeneous Spaces , 2018, ArXiv.

[21]  Koray Kavukcuoglu,et al.  Pixel Recurrent Neural Networks , 2016, ICML.

[22]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[23]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[24]  Andrew Zisserman,et al.  Spatial Transformer Networks , 2015, NIPS.

[25]  Rudrasis Chakraborty,et al.  Statistical Recurrent Models on Manifold valued Data , 2018, NIPS 2018.

[26]  Luc Van Gool,et al.  Building Deep Networks on Grassmann Manifolds , 2016, AAAI.

[27]  Rudrasis Chakraborty,et al.  An efficient recursive estimator of the Fréchet mean on a hypersphere with applications to Medical Image Analysis , 2015 .

[28]  Xavier Bresson,et al.  Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.

[29]  Silvere Bonnabel,et al.  Stochastic Gradient Descent on Riemannian Manifolds , 2011, IEEE Transactions on Automatic Control.

[30]  Stéphane Mallat,et al.  Understanding deep convolutional networks , 2016, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[31]  W. Kendall Probability, Convexity, and Harmonic Maps with Small Image I: Uniqueness and Fine Existence , 1990 .

[32]  N. Ayache,et al.  Log‐Euclidean metrics for fast and simple calculus on diffusion tensors , 2006, Magnetic resonance in medicine.

[33]  Luc Van Gool,et al.  A Riemannian Network for SPD Matrix Learning , 2016, AAAI.

[34]  M. Fréchet Les éléments aléatoires de nature quelconque dans un espace distancié , 1948 .

[35]  Mathieu Salzmann,et al.  Second-order Convolutional Neural Networks , 2017, ArXiv.

[36]  Stephan J. Garbin,et al.  Harmonic Networks: Deep Translation and Rotation Equivariance , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[37]  Jonathan H. Manton,et al.  A globally convergent numerical algorithm for computing the centre of mass on compact Lie groups , 2004, ICARCV 2004 8th Control, Automation, Robotics and Vision Conference, 2004..

[38]  Max Welling,et al.  Group Equivariant Convolutional Networks , 2016, ICML.

[39]  Nitish Srivastava,et al.  Unsupervised Learning of Video Representations using LSTMs , 2015, ICML.

[40]  Barnabás Póczos,et al.  Equivariance Through Parameter-Sharing , 2017, ICML.

[41]  Stéphane Mallat,et al.  Deep roto-translation scattering for object classification , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[42]  Barnabás Póczos,et al.  The Statistical Recurrent Unit , 2017, ICML.

[43]  Max Welling,et al.  Spherical CNNs , 2018, ICLR.

[44]  Christophe Lenglet,et al.  A nonparametric Riemannian framework for processing high angular resolution diffusion images and its applications to ODF-based morphometry , 2011, NeuroImage.

[45]  B. Afsari Riemannian Lp center of mass: existence, uniqueness, and convexity , 2011 .

[46]  Pierre Vandergheynst,et al.  Geometric Deep Learning: Going beyond Euclidean data , 2016, IEEE Signal Process. Mag..