The application of an oblique-projected Landweber method to a model of supervised learning

This paper brings together a novel information representation model for use in signal processing and computer vision problems, with a particular algorithmic development of the Landweber iterative algorithm. The information representation model allows a representation of multiple values for a variable as well as an expression for confidence. Both properties are important for effective computation using multi-level models, where a choice between models will be implementable as part of the optimization process. It is shown that in this way the algorithm can deal with a class of high-dimensional, sparse, and constrained least-squares problems, which arise in various computer vision learning tasks, such as object recognition and object pose estimation. While the algorithm has been applied to the solution of such problems, it has so far been used heuristically. In this paper we describe the properties and some of the peculiarities of the channel representation and optimization, and put them on firm mathematical ground. We consider the optimization a convexly constrained weighted least-squares problem and propose for its solution a projected Landweber method which employs oblique projections onto the closed convex constraint set. We formulate the problem, present the algorithm and work out its convergence properties, including a rate-of-convergence result. The results are put in perspective with currently available projected Landweber methods. An application to supervised learning is described, and the method is evaluated in an experiment involving function approximation, as well as application to transient signals.

[1]  Bertolt Eicke Iteration methods for convexly constrained ill-posed problems in hilbert space , 1992 .

[2]  Andrzej Stachurski,et al.  Parallel Optimization: Theory, Algorithms and Applications , 2000, Parallel Distributed Comput. Pract..

[3]  Björn Johansson,et al.  HiperLearn : A High Performance Learning Architecture , 2002 .

[4]  Gösta H. Granlund,et al.  Unrestricted Recognition of 3D Objects for Robotics Using Multilevel Triplet Invariants , 2004, AI Mag..

[5]  Anders Forsgren,et al.  Interior Methods for Nonlinear Optimization , 2002, SIAM Rev..

[6]  Boris Polyak,et al.  Constrained minimization methods , 1966 .

[7]  Paul Tseng,et al.  Error Bound and Reduced-Gradient Projection Algorithms for Convex Minimization over a Polyhedral Set , 1993, SIAM J. Optim..

[8]  Gösta H. Granlund,et al.  Unrestricted Recognition of 3-D Objects for Robotics Using Multi-Level Triplet Invariants , 2004 .

[9]  Thomas M. Cover,et al.  Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition , 1965, IEEE Trans. Electron. Comput..

[10]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[11]  S. Hyakin,et al.  Neural Networks: A Comprehensive Foundation , 1994 .

[12]  Y. Censor,et al.  Parallel Optimization:theory , 1997 .

[13]  Per-Erik Forssén Sparse Representations for Medium Level Vision , 2001 .

[14]  Per-Erik Forss,et al.  Sparse Representations for Medium Level Vision , 2001 .

[15]  Stephen J. Wright,et al.  Numerical Optimization , 2018, Fundamental Statistical Inference.

[16]  M. Bertero,et al.  Projected Landweber method and preconditioning , 1997 .

[17]  J. Navarro-Pedreño Numerical Methods for Least Squares Problems , 1996 .

[18]  Gösta H. Granlund,et al.  Unrestricted Recognition of 3-D Objects Using Multi-Level Triplet Invariants , 2002 .

[19]  Gunnar Rätsch,et al.  An introduction to kernel-based learning algorithms , 2001, IEEE Trans. Neural Networks.

[20]  Bart Kosko,et al.  Fuzzy Systems as Universal Approximators , 1994, IEEE Trans. Computers.

[21]  Mikael Adlers,et al.  Topics in Sparse Least Squares Problems , 2000 .

[22]  D. Bertsekas,et al.  Projection methods for variational inequalities with application to the traffic assignment problem , 1982 .

[23]  Robert J. Plemmons,et al.  Nonnegative Matrices in the Mathematical Sciences , 1979, Classics in Applied Mathematics.

[24]  K. Schittkowski,et al.  NONLINEAR PROGRAMMING , 2022 .

[25]  Gösta H. Granlund,et al.  An Associative Perception-Action Structure Using a Localized Space Variant Information Representation , 2000, AFPAC.