Applied Harmonic Analysis and Data Processing

Massive data sets have their own architecture. Each data source has an inherent structure, which we should attempt to detect in order to utilize it for applications, such as denoising, clustering, anomaly detection, knowledge extraction, or classification. Harmonic analysis revolves around creating new structures for decomposition, rearrangement and reconstruction of operators and functions—in other words inventing and exploring new architectures for information and inference. Two previous very successful workshops on applied harmonic analysis and sparse approximation have taken place in 2012 and in 2015. This workshop was the an evolution and continuation of these workshops and intended to bring together world leading experts in applied harmonic analysis, data analysis, optimization, statistics, and machine learning to report on recent developments, and to foster new developments and collaborations. Mathematics Subject Classification (2010): 42-XX, 65Txx, 94Axx, 65K05, 15A52. Introduction by the Organisers The workshop Applied Harmonic Analysis and Data Processing was organized by Ingrid Daubechies (Durham), Gitta Kutyniok (Berlin), Holger Rauhut (Aachen) and Thomas Strohmer (Davis). This meeting was attended by 49 participants from three continents. Advances in technology and the ever-growing role of digital sensors and computers in science have led to an exponential growth in the amount and complexity of data we collect. Uncertainty, scale, non-stationarity, noise, and heterogeneity are fundamental issues impeding progress at all phases of the pipeline that creates knowledge from data. This means that the amount of new 724 Oberwolfach Report 14/2018 mathematical challenges arising from the need of data analysis and information processing is enormous, with their solution requiring fundamentally new ideas and approaches, with significant consequences in the practical applications. Applied Harmonic Analysis provides a range of techniques towards the problem of efficiently representing, analyzing, compressing, and processing with “Big Data”. Massive data sets have their own architecture. Each data source has an inherent structure, which we should attempt to detect in order to utilize it for applications, such as denoising, clustering, anomaly detection, knowledge extraction, recovery, etc. Harmonic analysis revolves around creating new structures for decomposition, rearrangement and reconstruction of operators and functions—in other words inventing and exploring new architectures for information and inference. Indeed, in the last three decades Applied Harmonic Analysis has been at the center of many significant new ideas and methods crucial in a wide range of signal and image processing applications, and in the analysis and processing of large data sets. For example, compressive sensing, sparse approximations and models, geometric multiscale analysis and diffusion geometry represent some quite recent important breakthroughs. Several new directions have emerged on the heels of compressive sensing: Lowrank matrix recovery aims at recovering a matrix with small rank from incomplete data. In particular, matrix completion recovers the matrix from only a small fraction of its entries. Since low-rank structures arise in numerous applications, one can expect an enormous impact. However, much of the theory so far deals with linear measurements, while in practice we often also face non-linear measurements, for instance in situations where only signal intensity can be obtained. Despite recent breakthroughs in the area of phase retrieval, many challenging mathematical problems problems remain open in these areas. Inverse problems arising in connection with massive, complex data sets pose tremendous challenges and require new mathematical tools. Numerous deep questions arise. How can we utilize ideas of sparsity and minimal information complexity in this context? Is there a unified view of such measures that would include sparsity, lowrankness, and others (such as low-entropy), as special cases? This may lead to a new theory that considers an abstract notion of simplicity in general inverse problems. An important emerging topic in this context is the design efficient non-convex algorithms with provable convergence guarantees. One of the most exciting developments in machine learning in the past five years is the advent of deep learning, which is a special form of a neural network. Deep neural networks, and in particular convolutional networks have recently achieved state-of-the-art results on several complex computer vision and speech recognition tasks. However, until now deep learning acts very much like a black box, since algorithms are often based on ad hoc rules without theoretical foundation, the learned representations lack intepretability; we do not really understand why certain deep networks succeed and and we do not know how to modify them for those cases where they fail. Thus, developing a mathematical foundation for deep Applied Harmonic Analysis and Data Processing 725 learning is an important and rather challening task in data science, and one part of this workshop was dedicated to this topic. This workshop was a concerted effort to bring together researchers with various backgrounds, including harmonic analysis, optimization, probabibility theory, group theory, approximation theory, computer science, machine learning, and electrical engineering. The workshop featured 27 talks, thereof several longer overview talks. Moreover, a session of short presentations of 3 minutes took place on Monday, which we call the 3 Minutes of Fame (following Andy Warhols concept of 15 minutes of fame). This session has meanwhile become a tradition and has proven to be an efficient vehicle to ensure that every participant had the possibility to advertise her resarch. At the same time it is very entertaining for the audience. Almost all of the attendees participated, ranging from PhD students to renowned professors, contributing to the success of this session. Some highlights of the program included: • Advanced sampling theory: One of the problems that link harmonic analysis with data processing is the sampling problem. The main theoretical issue is how the stability of sampling and recovery is related to the number or density of samples. Related issues are the questions of localization, non-uniform sampling, and last not least suitable numerical algorithms. Karlheinz Gröchenig presented a range of compelling results using tools from shift-invariant spaces and totally positive functions. Albert Cohen discussed function approximation from sampling in high dimensions using optimal weighted least squares approximation. Felix Krahmer talked about “unlimited sampling”, a mathematical framework for sampling that can overcome limitations in current analog-to-digital converters. • Nonlinear inverse problems: In many applications we can only acquire nonlinear measurements of the function of interest. Phase retrieval is but the most prominent example. Several talks were dedicated to nonlinear inverse problems. Babak Hassibi and Rima Alaifari both presented recent progress in the solution of the phase retrieval problem, while Yuxin Chen and Justin Romberg highlighted exciting progress in convex and nonconvex optimization for certain nonlinear problems. • Emerging theory of Deep Learning: Despite the huge practical successes of Deep Learning in recent years, the mathematical understanding of deep learning is in its infancy. Several talks aimed at to remedy this situation. Philipp Grohs demonstrated how to avoid the curse of dimensionality when solving Kolmogorov equation in high dimensions by means of deep learning. Mahdi Soltanolkoltabi and Remi Gribonval were among several speakers who presented theoretical progress towards understanding some of the heuristics behind neural networks. The organizers would like to take the opportunity to thank MFO for providing support and a very inspiring environment for the workshop. The magic of the place and the pleasant atmosphere contributed greatly to the success of the workshop. 726 Oberwolfach Report 14/2018 Acknowledgement: The MFO and the workshop organizers would like to thank the National Science Foundation for supporting the participation of junior researchers in the workshop by the grant DMS-1641185, “US Junior Oberwolfach Fellows”. Applied Harmonic Analysis and Data Processing 727 Workshop: Applied Harmonic Analysis and Data Processing

[1]  J. Bahrdt,et al.  First observation of photons carrying orbital angular momentum in undulator radiation. , 2013, Physical review letters.

[2]  Christine Guenther,et al.  The fundamental solution on manifolds with time-dependent metrics , 2002 .

[3]  Amnon Shashua,et al.  SimNets: A Generalization of Convolutional Networks , 2014, ArXiv.

[4]  K. Gröchenig,et al.  Sharp Results on Sampling with Derivatives in Shift-Invariant Spaces and Multi-Window Gabor Frames , 2017, Constructive Approximation.

[5]  D. Gottesman An Introduction to Quantum Error Correction and Fault-Tolerant Quantum Computation , 2009, 0904.2557.

[6]  Mark M. Wilde,et al.  Logical operators of quantum codes , 2009, 0903.5256.

[7]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[8]  J. Romberg,et al.  A flexible convex relaxation for phase retrieval , 2017 .

[9]  Simon Foucart,et al.  Flavors of Compressive Sensing , 2016 .

[10]  Matus Telgarsky,et al.  Benefits of Depth in Neural Networks , 2016, COLT.

[11]  Hrushikesh Narhar Mhaskar,et al.  Approximation properties of a multilayered feedforward artificial neural network , 1993, Adv. Comput. Math..

[12]  Sanjeev Arora,et al.  New Algorithms for Learning Incoherent and Overcomplete Dictionaries , 2013, COLT.

[13]  Akram Aldroubi,et al.  Nonuniform Sampling and Reconstruction in Shift-Invariant Spaces , 2001, SIAM Rev..

[14]  Peter Jung,et al.  Blind sparse recovery from superimposed non-linear sensor measurements , 2017, 2017 International Conference on Sampling Theory and Applications (SampTA).

[15]  Rémi Gribonval,et al.  Flexible Multilayer Sparse Approximations of Matrices and Applications , 2016, IEEE Journal of Selected Topics in Signal Processing.

[16]  Roy R. Lederman,et al.  Common Manifold Learning Using Alternating-Diffusion , 2015 .

[17]  Gitta Kutyniok,et al.  A Mathematical Framework for Feature Selection from Real-World Data with Non-Linear Observations , 2016, ArXiv.

[18]  Yaniv Plan,et al.  Robust 1-bit Compressed Sensing and Sparse Logistic Regression: A Convex Programming Approach , 2012, IEEE Transactions on Information Theory.

[19]  R. Bhatia Positive Definite Matrices , 2007 .

[20]  Ronen Tamari,et al.  Boosting Dilated Convolutional Networks with Mixed Tensor Decompositions , 2017, ICLR.

[21]  L. Rudin,et al.  Nonlinear total variation based noise removal algorithms , 1992 .

[22]  Andrew R. Barron,et al.  Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.

[23]  Ingrid Daubechies,et al.  Reconstructing Real-Valued Functions from Unsigned Coefficients with Respect to Wavelet and Other Frames , 2016, 1601.07579.

[24]  Sayan Mukherjee,et al.  The Geometry of Synchronization Problems and Learning Group Actions , 2016, Discrete & Computational Geometry.

[25]  Suvrit Sra,et al.  Conic Geometric Optimization on the Manifold of Positive Definite Matrices , 2013, SIAM J. Optim..

[26]  Hrushikesh Narhar Mhaskar,et al.  Weighted quadrature formulas and approximation by zonal function networks on the sphere , 2006, J. Complex..

[27]  Dmitry Yarotsky,et al.  Error bounds for approximations with deep ReLU networks , 2016, Neural Networks.

[28]  John Wright,et al.  Complete Dictionary Recovery Over the Sphere II: Recovery by Riemannian Trust-Region Method , 2015, IEEE Transactions on Information Theory.

[29]  Dustin G. Mixon,et al.  Monte Carlo approximation certificates for k-means clustering , 2017, 1710.00956.

[30]  Felipe Cucker,et al.  On the mathematical foundations of learning , 2001 .

[31]  Richard G. Baraniuk,et al.  Exponential Decay of Reconstruction Error From Binary Measurements of Sparse Signals , 2014, IEEE Transactions on Information Theory.

[32]  C. Chui,et al.  Article in Press Applied and Computational Harmonic Analysis a Randomized Algorithm for the Decomposition of Matrices , 2022 .

[33]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[34]  Bragg-von Laue diffraction generalized to twisted X-rays. , 2016, Acta crystallographica. Section A, Foundations and advances.

[35]  Dustin G. Mixon,et al.  Probably certifiably correct k-means clustering , 2015, Math. Program..

[36]  Rui Chao,et al.  Fault-tolerant quantum computation with few qubits , 2017, npj Quantum Information.

[37]  Matus Telgarsky,et al.  Representation Benefits of Deep Feedforward Networks , 2015, ArXiv.

[38]  Mahdi Soltanolkotabi,et al.  Learning ReLUs via Gradient Descent , 2017, NIPS.

[39]  Philipp Grohs,et al.  Phase Retrieval In The General Setting Of Continuous Frames For Banach Spaces , 2016, SIAM J. Math. Anal..

[40]  José Luis Romero,et al.  Sampling theorems for shift-invariant spaces, Gabor frames, and totally positive functions , 2016, Inventiones mathematicae.

[41]  Jiming Peng,et al.  Advanced Optimization Laboratory Title : Approximating K-means-type clustering via semidefinite programming , 2005 .

[42]  Bernhard Schölkopf,et al.  A Kernel Two-Sample Test , 2012, J. Mach. Learn. Res..

[43]  Philipp Petersen,et al.  Optimal approximation of piecewise smooth functions using deep ReLU neural networks , 2017, Neural Networks.

[44]  Martin Genzel,et al.  High-Dimensional Estimation of Structured Signals From Non-Linear Observations With General Convex Loss Functions , 2016, IEEE Transactions on Information Theory.

[45]  Albert Cohen,et al.  Sequential Sampling for Optimal Weighted Least Squares Approximations in Hierarchical Spaces , 2018, SIAM J. Math. Data Sci..

[46]  Ramesh Raskar,et al.  Unlimited Sampling of Sparse Signals , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[47]  Amnon Shashua,et al.  Inductive Bias of Deep Convolutional Networks through Pooling Geometry , 2016, ICLR.

[48]  K. Gröchenig The Mystery of Gabor Frames , 2014 .

[49]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[50]  Gabriele Steidl,et al.  Examplar-Based Face Colorization Using Image Morphing , 2017, J. Imaging.

[51]  Sjoerd Dirksen,et al.  Robust one-bit compressed sensing with non-Gaussian measurements , 2018, ArXiv.

[52]  Adel Javanmard,et al.  Theoretical Insights Into the Optimization Landscape of Over-Parameterized Shallow Neural Networks , 2017, IEEE Transactions on Information Theory.

[53]  Shor,et al.  Good quantum error-correcting codes exist. , 1995, Physical review. A, Atomic, molecular, and optical physics.

[54]  Benjamin Berkels,et al.  Time Discrete Geodesic Paths in the Space of Images , 2015, SIAM J. Imaging Sci..

[55]  Justin Romberg,et al.  Solving Equations of Random Convex Functions via Anchored Regression , 2017, Found. Comput. Math..

[56]  Dominik Jüstel,et al.  Twisted X-Rays: Incoming Waveforms Yielding Discrete Diffraction Patterns for Helical Structures , 2015, SIAM J. Appl. Math..

[57]  Peter G. Casazza,et al.  Phase retrieval in infinite-dimensional Hilbert spaces , 2016, 1601.06411.

[58]  Alexandros G. Dimakis,et al.  Compressed Sensing using Generative Models , 2017, ICML.

[59]  Matthew J. Hirn,et al.  Time Coupled Diffusion Maps , 2016, Applied and Computational Harmonic Analysis.

[60]  Tao Zhou,et al.  A Christoffel function weighted least squares algorithm for collocation approximations , 2014, Math. Comput..

[61]  John Wright,et al.  Complete Dictionary Recovery Over the Sphere I: Overview and the Geometric Picture , 2015, IEEE Transactions on Information Theory.

[62]  D. Gottesman Theory of fault-tolerant quantum computation , 1997, quant-ph/9702029.

[63]  Holger Rauhut,et al.  Low rank matrix recovery from rank one measurements , 2014, ArXiv.

[64]  Mitsuo Morimoto Analytic functionals on the sphere and their Fourier-Borel transformations , 1983 .

[65]  Karin Schnass,et al.  Convergence radius and sample complexity of ITKM algorithms for dictionary learning , 2015, Applied and Computational Harmonic Analysis.

[66]  Xavier Pennec,et al.  A Riemannian Framework for Tensor Computing , 2005, International Journal of Computer Vision.

[67]  M. Elad,et al.  $rm K$-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation , 2006, IEEE Transactions on Signal Processing.

[68]  Gabriele Steidl,et al.  Morphing of Manifold-Valued Images Inspired by Discrete Geodesics in Image Spaces , 2017, SIAM J. Imaging Sci..

[69]  Dustin G. Mixon,et al.  Saving phase: Injectivity and stability for phase retrieval , 2013, 1302.4618.

[70]  Yaniv Plan,et al.  The Generalized Lasso With Non-Linear Observations , 2015, IEEE Transactions on Information Theory.

[71]  Amin Karbasi,et al.  Gradient Methods for Submodular Maximization , 2017, NIPS.

[72]  Quoc Thong Le Gia,et al.  Localized Linear Polynomial Operators and Quadrature Formulas on the Sphere , 2008, SIAM J. Numer. Anal..

[73]  Tingran Gao,et al.  Semi‐supervised determination of pseudocryptic morphotypes using observer‐free characterizations of anatomical alignment and shape , 2017, Ecology and evolution.

[74]  Anima Anandkumar,et al.  Exact Recovery of Sparsely Used Overcomplete Dictionaries , 2013, ArXiv.

[75]  Steane,et al.  Simple quantum error-correcting codes. , 1996, Physical review. A, Atomic, molecular, and optical physics.

[76]  Peter Jung,et al.  Recovering Structured Data From Superimposed Non-Linear Measurements , 2017, IEEE Transactions on Information Theory.

[77]  E. Candès,et al.  Stable signal recovery from incomplete and inaccurate measurements , 2005, math/0503066.

[78]  Justin K. Romberg,et al.  Compressive Sensing by Random Convolution , 2009, SIAM J. Imaging Sci..

[79]  Allan Pinkus,et al.  Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.

[80]  A. Cohen,et al.  Optimal weighted least-squares methods , 2016, 1608.00512.

[81]  Pierre Weiss,et al.  Compressed sensing with structured sparsity and structured acquisition , 2015, Applied and Computational Harmonic Analysis.

[82]  N. J. A. Sloane,et al.  Quantum Error Correction Via Codes Over GF(4) , 1998, IEEE Trans. Inf. Theory.

[83]  Dustin G. Mixon,et al.  Clustering subgaussian mixtures by semidefinite programming , 2016, ArXiv.

[84]  Dominik Jüstel,et al.  The Zak transform on strongly proper G ‐spaces and its applications , 2018, J. Lond. Math. Soc..

[85]  Stéphane Lafon,et al.  Diffusion maps , 2006 .

[86]  Martin Roetteler,et al.  Leveraging automorphisms of quantum codes for fault-tolerant quantum computation , 2013, 2013 IEEE International Symposium on Information Theory.

[87]  Christian Jutten,et al.  Classification of covariance matrices using a Riemannian-based kernel for BCI applications , 2013, Neurocomputing.

[88]  Ramesh Raskar,et al.  Unlimited Sampling of Sparse Sinusoidal Mixtures , 2018, 2018 IEEE International Symposium on Information Theory (ISIT).

[89]  Rayan Saab,et al.  One-Bit Compressive Sensing With Norm Estimation , 2014, IEEE Transactions on Information Theory.

[90]  A. Robert Calderbank,et al.  Synthesis of Logical Clifford Operators via Symplectic Geometry , 2018, 2018 IEEE International Symposium on Information Theory (ISIT).

[91]  Felipe Cucker,et al.  Learning Theory: An Approximation Theory Viewpoint: References , 2007 .

[92]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[93]  I. Johnstone,et al.  Ideal spatial adaptation by wavelet shrinkage , 1994 .

[94]  Alireza Doostan,et al.  Coherence motivated sampling and convergence analysis of least squares polynomial Chaos regression , 2014, 1410.1931.

[95]  Meena Mahajan,et al.  The Planar k-means Problem is NP-hard I , 2009 .

[96]  Yaniv Plan,et al.  One-bit compressed sensing with non-Gaussian measurements , 2012, ArXiv.

[97]  Ohad Shamir,et al.  Depth-Width Tradeoffs in Approximating Natural Functions with Neural Networks , 2016, ICML.

[98]  Vladislav Voroninski,et al.  Global Guarantees for Enforcing Deep Generative Priors by Empirical Risk , 2017, IEEE Transactions on Information Theory.

[99]  Tingran Gao,et al.  Hypoelliptic Diffusion Maps and Their Applications in Automated Geometric Morphometrics , 2015 .

[100]  Sebastian Becker,et al.  Solving stochastic differential equations and Kolmogorov equations by means of deep learning , 2018, ArXiv.

[101]  Richard G. Baraniuk,et al.  1-Bit compressive sensing , 2008, 2008 42nd Annual Conference on Information Sciences and Systems.

[102]  Hrushikesh Narhar Mhaskar,et al.  When is approximation by Gaussian networks necessarily a linear process? , 2004, Neural Networks.

[103]  David Steurer,et al.  Dictionary Learning and Tensor Decomposition via the Sum-of-Squares Method , 2014, STOC.

[104]  David Berthelot,et al.  BEGAN: Boundary Equilibrium Generative Adversarial Networks , 2017, ArXiv.