To accelerate the existing Broad Learning System (BLS) for new added nodes in [7], we extend the inverse Cholesky factorization in [10] to deduce an efficient inverse Cholesky factorization for a Hermitian matrix partitioned into 2 * 2 blocks, which is utilized to develop the proposed BLS algorithm 1. The proposed BLS algorithm 1 compute the ridge solution (i.e, the output weights) from the inverse Cholesky factor of the Hermitian matrix in the ridge inverse, and update the inverse Cholesky factor efficiently. From the proposed BLS algorithm 1, we deduce the proposed ridge inverse, which can be obtained from the generalized inverse in [7] by just change one matrix in the equation to compute the newly added sub-matrix. We also modify the proposed algorithm 1 into the proposed algorithm 2, which is equivalent to the existing BLS algorithm [7] in terms of numerical computations. The proposed algorithms 1 and 2 can reduce the computational complexity, since usually the Hermitian matrix in the ridge inverse is smaller than the ridge inverse. With respect to the existing BLS algorithm, the proposed algorithms 1 and 2 usually require about 13 and 2 3 of complexities, respectively, while in numerical experiments they achieve the speedups (in each additional training time) of 2.40 - 2.91 and 1.36 - 1.60, respectively. Numerical experiments also show that the proposed algorithm 1 and the standard ridge solution always bear the same testing accuracy, and usually so do the proposed algorithm 2 and the existing BLS algorithm. The existing BLS assumes the ridge parameter lamda->0, since it is based on the generalized inverse with the ridge regression approximation. When the assumption of lamda-> 0 is not satisfied, the standard ridge solution obviously achieves a better testing accuracy than the existing BLS algorithm in numerical experiments.
[1]
Isabelle Guyon,et al.
Neural Network Recognizer for Hand-Written Zip Code Digits
,
1988,
NIPS.
[2]
Dejan J. Sobajic,et al.
Learning and generalization characteristics of the random vector Functional-link net
,
1994,
Neurocomputing.
[3]
S. R. Searle,et al.
On Deriving the Inverse of a Sum of Matrices
,
1981
.
[4]
Allan Pinkus,et al.
Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function
,
1991,
Neural Networks.
[5]
David M. Auslander,et al.
Control and dynamic systems
,
1970
.
[6]
Gene H. Golub,et al.
Matrix computations (3rd ed.)
,
1996
.
[7]
C. L. Philip Chen,et al.
A rapid learning and dynamic stepwise updating algorithm for flat neural networks and the application to time-series prediction
,
1999,
IEEE Trans. Syst. Man Cybern. Part B.
[8]
Bin Li,et al.
An Improved Square-Root Algorithm for V-BLAST Based on Efficient Inverse Cholesky Factorization
,
2020,
IEEE Transactions on Wireless Communications.
[9]
Yoshua Bengio,et al.
Gradient-based learning applied to document recognition
,
1998,
Proc. IEEE.
[10]
Donald W. Marquaridt.
Generalized Inverses, Ridge Regression, Biased Linear Estimation, and Nonlinear Estimation
,
1970
.
[11]
C. L. Philip Chen,et al.
Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture
,
2018,
IEEE Transactions on Neural Networks and Learning Systems.
[12]
C. L. Philip Chen,et al.
Universal Approximation Capability of Broad Learning System and Its Structural Variations
,
2019,
IEEE Transactions on Neural Networks and Learning Systems.
[13]
Lawrence D. Jackel,et al.
Handwritten Digit Recognition with a Back-Propagation Network
,
1989,
NIPS.