Signs of divided differences yield least squares data fitting with constrained monotonicity or convexity

Methods are presented for least squares data smoothing by using the signs of divided differences of the smoothed values. Professor M.J.D. Powell initiated the subject in the early 1980s and since then, theory, algorithms and FORTRAN software make it applicable to several disciplines in various ways.Let us consider n data measurements of a univariate function which have been altered by random errors. Then it is usual for the divided differences of the measurements to show sign alterations, which are probably due to data errors. We make the least sum of squares change to the measurements, by requiring the sequence of divided differences of order m to have at most q sign changes for some prescribed integer q. The positions of the sign changes are integer variables of the optimization calculation, which implies a combinatorial problem whose solution can require about O(nq) quadratic programming calculations in n variables and n - m constraints.Suitable methods have been developed for the following cases. It has been found that a dynamic programming procedure can calculate the global minimum for the important cases of piecewise monotonicity m = 1, q ≥ 1 and piecewise convexity/concavity m = 2, q ≥ 1 of the smoothed values. The complexity of the procedure in the case of m = 1 is O(n2 + qn log2 n) computer operations, while it is reduced to only O(n) when q = 0 (monotonicity) and q = 1 (increasing/decreasing monotonicity). The case m = 2, q ≥ 1 requires O(qn2) computer operations and n2 quadratic programming calculations, which is reduced to one and n - 2 quadratic programming calculations when m = 2, q = 0, i.e. convexity, and m = 2, q = 1, i.e. convexity/concavity, respectively.Unfortunately, the technique that receives this efficiency cannot generalize for the highly nonlinear case m ≥ 3,q ≥ 2. However, the case m ≥ 3,q = 0 is solved by a special strictly convex quadratic programming calculation, and the case m ≥ 3,q = 1 can be solved by at most 2(n - m) applications of the previous algorithm. Also, as m gets larger, large sets of active constraints usually occur at the optimal approximation, which makes the calculation for higher values of q less expensive than what it seems to. Further, attention is given to the sensitivity of the solution with respect to changes in the constraints and the data.The smoothing technique is an active research topic and there seems to be room for further developments. One strong reason for studying methods that make use of divided differences for data smoothing is that, whereas only the data are provided, the achieved accuracy goes much beyond the accuracy of the data at an order determined automatically by the chosen divided differences.

[1]  C. Lanczos Applied Analysis , 1961 .

[2]  Wenceslao González-Manteiga,et al.  Data-Driven Smoothing Based on Convexity Properties , 1991 .

[3]  J. Rice Bandwidth Choice for Nonparametric Regression , 1984 .

[4]  I. C. Demetriou CXFTV2: A Fortran subroutine for the discrete least squares convex approximation , 1997 .

[5]  Alexander Rosenberg,et al.  Economics--Mathematical Politics or Science of Diminishing Returns? , 1994 .

[6]  Related Topics,et al.  Nonparametric functional estimation and related topics , 1991 .

[7]  Constance Van Eeden Maximum Likelihood Estimation Of Ordered Probabilities1) , 1956 .

[8]  W. Härdle Applied Nonparametric Regression , 1991 .

[9]  W. Edwards,et al.  Decision Analysis and Behavioral Research , 1986 .

[10]  W. Härdle Applied Nonparametric Regression , 1992 .

[11]  R. Fletcher Practical Methods of Optimization , 1988 .

[12]  C. Hildreth Point Estimates of Ordinates of Concave Functions , 1954 .

[13]  M. P. Cullinan Data smoothing using non-negative divided differences and l2 approximation , 1990 .

[14]  Data smoothing by divided differences , 1982 .

[15]  Wu Li,et al.  A Data Smoothing Technique for Piecewise Convex/Concave Curves , 1996, SIAM J. Sci. Comput..

[16]  Arieh Iserles,et al.  Approximation theory and optimization - tributes to M. J. D. Powell , 2009 .

[17]  I. C. Demetriou,et al.  Certain positive definite submatrices that arise from binomial coefficient matrices , 2001 .

[18]  Paul Dierckx,et al.  Curve and surface fitting with splines , 1994, Monographs on numerical analysis.

[19]  J. Kalbfleisch Statistical Inference Under Order Restrictions , 1975 .

[20]  E. Mammen Nonparametric regression under qualitative smoothness assumptions , 1991 .

[21]  Harold J. Kushner,et al.  A New Method of Locating the Maximum Point of an Arbitrary Multipeak Curve in the Presence of Noise , 1964 .

[22]  Ioannis C. Demetriou,et al.  Algorithm 742: L2CXFT: a Fortran subroutine for least-squares data fitting with nonnegative second divided differences , 1995, TOMS.

[23]  J. Kruskal Nonmetric multidimensional scaling: A numerical method , 1964 .

[24]  Carl de Boor,et al.  A Practical Guide to Splines , 1978, Applied Mathematical Sciences.

[25]  I. C. Demetriou Discrete piecewise monotonic approximation by a strictly convex distance function , 1995 .

[26]  L. Schumaker Spline Functions: Basic Theory , 1981 .

[27]  Paul Wintz,et al.  Digital image processing (2nd ed.) , 1987 .

[28]  Donald Goldfarb,et al.  A numerically stable dual method for solving strictly convex quadratic programs , 1983, Math. Program..

[29]  Gene H. Golub,et al.  Matrix computations , 1983 .

[30]  M. J. D. Powell,et al.  Least Squares Smoothing of Univariate Data to achieve Piecewise Monotonicity , 1991 .

[31]  M. Powell,et al.  Approximation theory and methods , 1984 .

[32]  F. T. Wright,et al.  Order restricted statistical inference , 1988 .

[33]  M. J. D. Powell,et al.  The Minimum Sum of Squares Change to Univariate Data that gives Convexity , 1991 .