The global minimum search problem is important in neural networks because the error cost involved is formed as multiminima potential in weight parametric space. Therefore, parameters that produce a global minimum in a cost function are the best values for enhancing the performance of neural networks. Previously, a global minimum search based on a damped oscillator equation known as the heavy ball with friction (HBF) was studied. The kinetic energy overcomes a local minimum if the kinetic energy is sufficiently large or else the heavy ball will converge into a local minimum due to the action of friction. However, an appropriate damping coefficient has not been found in the HBF; therefore, the ball has to be shot again after it arrives at each local minimum until it finds a global minimum. In order to solve this problem, we determined an adaptive damping coefficient using the geodesic of Newtonian dynamics Lagrangian. This geometric method produces a second-order adaptively damped oscillator equation, the damping coefficient of which is the negative time derivative of the logarithmic function of the cost potential. Furthermore, we obtained a novel adaptive steepest descent by discretizing this second-order equation. To investigate the performance of this novel steepest descent, we applied our first-order update rule to the Rosenbrock- and Griewank-type potentials. The results show that our method determined the global minimum in most cases from various initial points. Our adaptive steepest descent may be applied in many fields related to global minimum search, such as neural networks, game theory, and economics.
[1]
Shotaro Akaho,et al.
Learning algorithms utilizing quasi-geodesic flows on the Stiefel manifold
,
2005,
Neurocomputing.
[2]
S Amelinckx,et al.
Classical Dynamics of Particles and Systems
,
1971
.
[3]
J. Gillis,et al.
Classical dynamics of particles and systems
,
1965
.
[4]
H. Attouch,et al.
THE HEAVY BALL WITH FRICTION METHOD, I. THE CONTINUOUS DYNAMICAL SYSTEM: GLOBAL EXPLORATION OF THE LOCAL MINIMA OF A REAL-VALUED FUNCTION BY ASYMPTOTIC ANALYSIS OF A DISSIPATIVE DYNAMICAL SYSTEM
,
2000
.
[5]
Alan Edelman,et al.
The Geometry of Algorithms with Orthogonality Constraints
,
1998,
SIAM J. Matrix Anal. Appl..
[6]
A. Cabot.
Inertial Gradient-Like Dynamical System Controlled by a Stabilizing Term
,
2004
.
[7]
JOHN L. Martin.
General Relativity: A First Course for Physicists
,
1995
.
[8]
Ning Qian,et al.
On the momentum term in gradient descent learning algorithms
,
1999,
Neural Networks.
[9]
Tommy W. S. Chow,et al.
Training multilayer neural networks using fast global learning algorithm - least-squares and penalized optimization methods
,
1999,
Neurocomputing.
[10]
C. Clementi,et al.
GEOMETRY OF DYNAMICS, LYAPUNOV EXPONENTS, AND PHASE TRANSITIONS
,
1997,
chao-dyn/9702011.