Deep connections between learning from limited labels & physical parameter estimation - inspiration for regularization

Recently established equivalences between differential equations and the structure of neural networks enabled some interpretation of training of a neural network as partial-differential-equation (PDE) constrained optimization. We add to the previously established connections, explicit regularization that is particularly beneficial in the case of single large-scale examples with partial annotation. We show that explicit regularization of model parameters in PDE constrained optimization translates to regularization of the network output. Examination of the structure of the corresponding Lagrangian and backpropagation algorithm do not reveal additional computational challenges. A hyperspectral imaging example shows that minimum prior information together with cross-validation for optimal regularization parameters boosts the segmentation accuracy.

[1]  Eldad Haber,et al.  Does shallow geological knowledge help neural-networks to predict deep units? , 2019 .

[2]  Eldad Haber,et al.  Deep Neural Networks Motivated by Partial Differential Equations , 2018, Journal of Mathematical Imaging and Vision.

[3]  Zhixiang Xue,et al.  A general generative adversarial capsule network for hyperspectral image spectral-spatial classification , 2020, Remote Sensing Letters.

[4]  Ying Li,et al.  Spectral-Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network , 2017, Remote. Sens..

[5]  Zheng Xu,et al.  Training Neural Networks Without Gradients: A Scalable ADMM Approach , 2016, ICML.

[6]  Heesung Kwon,et al.  Contextual deep CNN based hyperspectral classification , 2016, 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS).

[7]  Lars Ruthotto,et al.  Layer-Parallel Training of Deep Residual Neural Networks , 2018, SIAM J. Math. Data Sci..

[8]  Yann Le Cun,et al.  A Theoretical Framework for Back-Propagation , 1988 .

[9]  Eran Treister,et al.  IMEXnet: A Forward Stable Deep Neural Network , 2019, ICML.

[10]  Kurt Keutzer,et al.  ANODEV2: A Coupled Neural ODE Evolution Framework , 2019, ArXiv.

[11]  Mahdi Hasanlou,et al.  Hyperspectral change detection: an experimental comparative study , 2018 .

[12]  Eldad Haber,et al.  Reversible Architectures for Arbitrarily Deep Residual Neural Networks , 2017, AAAI.

[13]  Ghassan Hamarneh,et al.  Topology Aware Fully Convolutional Networks for Histology Gland Segmentation , 2016, MICCAI.

[14]  Eldad Haber,et al.  Neural-networks for geophysicists and their application to seismic data interpretation , 2019, The Leading Edge.

[15]  David Duvenaud,et al.  Neural Ordinary Differential Equations , 2018, NeurIPS.

[16]  Kurt Keutzer,et al.  ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs , 2019, IJCAI.

[17]  Keegan Lensink,et al.  Symmetric block-low-rank layers for fully reversible multilevel neural networks , 2019, ArXiv.

[18]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Bo Li,et al.  Multi-scale 3D deep convolutional neural network for hyperspectral image classification , 2017, 2017 IEEE International Conference on Image Processing (ICIP).

[20]  Eldad Haber,et al.  Stable architectures for deep neural networks , 2017, ArXiv.

[21]  Yong Xiao,et al.  CSA-MSO3DCNN: Multiscale Octave 3D CNN with Channel and Spatial Attention for Hyperspectral Image Classification , 2020, Remote. Sens..

[22]  Eran Treister,et al.  LeanConvNets: Low-Cost Yet Effective Convolutional Neural Networks , 2020, IEEE Journal of Selected Topics in Signal Processing.

[23]  Daniel Cremers,et al.  Regularization for Deep Learning: A Taxonomy , 2017, ArXiv.

[24]  Eldad Haber,et al.  Fully hyperbolic convolutional neural networks , 2019, Research in the Mathematical Sciences.

[25]  Eldad Haber,et al.  Multi-resolution neural networks for tracking seismic horizons from few training images , 2018, Interpretation.