Multilayer Perceptrons: Other Learning Techniques
暂无分享,去创建一个
[1] Kenji Fukumizu,et al. Adaptive Method of Realizing Natural Gradient Learning for Multilayer Perceptrons , 2000, Neural Computation.
[2] Samir Kouro,et al. Unidimensional Modulation Technique for Cascaded Multilevel Converters , 2009, IEEE Transactions on Industrial Electronics.
[3] Hao Yu,et al. Neural Network Learning Without Backpropagation , 2010, IEEE Transactions on Neural Networks.
[4] Andrew Chi-Sing Leung,et al. On the Kalman filtering method in neural network training and pruning , 1999, IEEE Trans. Neural Networks.
[5] Etienne Barnard,et al. Optimization for training neural nets , 1992, IEEE Trans. Neural Networks.
[6] John L. Nazareth. Differentiable optimization and equation solving , 2003 .
[7] Michael T. Manry,et al. A neural network training algorithm utilizing multiple sets of linear equations , 1996, Conference Record of The Thirtieth Asilomar Conference on Signals, Systems and Computers.
[8] Tom Heskes,et al. On Natural Learning and Pruning in Multilayered Perceptrons , 2000, Neural Computation.
[9] Antanas Verikas,et al. Training neural networks by stochastic optimisation , 2000, Neurocomputing.
[10] Francesco Piazza,et al. On the complex backpropagation algorithm , 1992, IEEE Trans. Signal Process..
[11] Francesco Palmieri,et al. MEKA-a fast, local algorithm for training feedforward neural networks , 1990, 1990 IJCNN International Joint Conference on Neural Networks.
[12] M. N. Shanmukha Swamy,et al. Complex EKF neural network for adaptive equalization , 2000, 2000 IEEE International Symposium on Circuits and Systems. Emerging Technologies for the 21st Century. Proceedings (IEEE Cat No.00CH36353).
[13] Robert M. Burton,et al. Event-dependent control of noise enhances learning in neural networks , 1992, Neural Networks.
[14] Hao Yu,et al. Improved Computation for Levenberg–Marquardt Training , 2010, IEEE Transactions on Neural Networks.
[15] Chin-Teng Lin,et al. A second-order learning algorithm for multilayer networks based on block Hessian matrix , 1998, Neural Networks.
[16] Chao Zhang,et al. Convergence of Batch Split-Complex Backpropagation Algorithm for Complex-Valued Neural Networks , 2009 .
[17] Sammy Siu,et al. Sensitivity Analysis of the Split-Complex Valued Multilayer Perceptron Due to the Errors of the i.i.d. Inputs and Weights , 2007, IEEE Transactions on Neural Networks.
[18] Stavros J. Perantonis,et al. Two highly efficient second-order algorithms for training feedforward networks , 2002, IEEE Trans. Neural Networks.
[19] Kwok-Wo Wong,et al. Generalized RLS approach to the training of neural networks , 2006, IEEE Trans. Neural Networks.
[20] Cris Koutsougeras,et al. Complex domain backpropagation , 1992 .
[21] Hideaki Sakai,et al. A real-time learning algorithm for a multilayered neural network based on the extended Kalman filter , 1992, IEEE Trans. Signal Process..
[22] Jaewook Lee. Attractor-based trust-region algorithm for efficient training of multilayer perceptrons , 2003 .
[23] Léon Personnaz,et al. A recursive algorithm based on the extended Kalman filter for the training of feedforward neural models , 1998, Neurocomputing.
[24] Stavros J. Perantonis,et al. Training feedforward neural networks with the Dogleg method and BFGS Hessian updates , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.
[25] Nikolai S. Rubanov. The layer-wise method and the backpropagation hybrid approach to learning a feedforward neural network , 2000, IEEE Trans. Neural Networks Learn. Syst..
[26] H.S.M. Beigi. Neural network learning through optimally conditioned quadratically convergent methods requiring NO LINE SEARCH , 1993, Proceedings of 36th Midwest Symposium on Circuits and Systems.
[27] Mark E. Oxley,et al. Comparative Analysis of Backpropagation and the Extended Kalman Filter for Training Multilayer Perceptrons , 1992, IEEE Trans. Pattern Anal. Mach. Intell..
[28] Sharad Singhal,et al. Training feed-forward networks with the extended Kalman algorithm , 1989, International Conference on Acoustics, Speech, and Signal Processing,.
[29] Danilo P. Mandic,et al. A normalised complex backpropagation algorithm , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.
[30] Paramasivan Saratchandran,et al. A new learning algorithm with logarithmic performance index for complex-valued neural networks , 2009, Neurocomputing.
[31] Daesik Hong,et al. Nonlinear blind equalization schemes using complex-valued multilayer feedforward neural networks , 1998, IEEE Trans. Neural Networks.
[32] Gregory J. Wolff,et al. Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.
[33] Stefano Fanelli,et al. A new class of quasi-Newtonian methods for optimal learning in MLP-networks , 2003, IEEE Trans. Neural Networks.
[34] George W. Irwin,et al. Fast parallel off-line training of multilayer perceptrons , 1997, IEEE Trans. Neural Networks.
[35] Michael Fairbank,et al. Efficient Calculation of the Gauss-Newton Approximation of the Hessian Matrix in Neural Networks , 2012, Neural Computation.
[36] Andrew Luk,et al. Fast Convergent Generalized Back-Propagation Algorithm with Constant Learning Rate , 2004, Neural Processing Letters.
[37] Okyay Kaynak,et al. Computing Gradient Vector and Jacobian Matrix in Arbitrarily Connected Neural Networks , 2008, IEEE Transactions on Industrial Electronics.
[38] Vijanth S. Asirvadam,et al. A memory optimal BFGS neural network training algorithm , 2002, Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290).
[39] Leszek Rutkowski,et al. A fast training algorithm for neural networks , 1998 .
[40] Chris Bishop,et al. Exact Calculation of the Hessian Matrix for the Multilayer Perceptron , 1992, Neural Computation.
[41] Sagar V. Kamarthi,et al. Accelerating neural network training using weight extrapolations , 1999, Neural Networks.
[42] Roberto Battiti,et al. First- and Second-Order Methods for Learning: Between Steepest Descent and Newton's Method , 1992, Neural Computation.
[43] José R. Dorronsoro,et al. Natural conjugate gradient training of multilayer perceptrons , 2006, Neurocomputing.
[44] A. Perry. A Modified Conjugate Gradient Algorithm for Unconstrained Nonlinear Optimization , 1975 .
[45] R. Fletcher. Practical Methods of Optimization , 1988 .
[46] David Zhang,et al. Parameter by Parameter Algorithm for Multilayer Perceptrons , 2006, Neural Processing Letters.
[47] Paolo Campolucci,et al. Complex-valued neural networks with adaptive spline activation function for digital-radio-links nonlinear equalization , 1999, IEEE Trans. Signal Process..
[48] Eugenius Kaszkurewicz,et al. Steepest descent with momentum for quadratic functions is a version of the conjugate gradient method , 2004, Neural Networks.
[49] M. Hestenes,et al. Methods of conjugate gradients for solving linear systems , 1952 .
[50] Tülay Adali,et al. Approximation by Fully Complex Multilayer Perceptrons , 2003, Neural Computation.
[51] C. M. Reeves,et al. Function minimization by conjugate gradients , 1964, Comput. J..
[52] Jorge J. Moré,et al. The Levenberg-Marquardt algo-rithm: Implementation and theory , 1977 .
[53] George W. Irwin,et al. A Variable Memory Quasi-Newton Training Algorithm , 1999, Neural Processing Letters.
[54] Avinoam Perry,et al. Technical Note - A Modified Conjugate Gradient Algorithm , 1978, Oper. Res..
[55] Michael T. Manry,et al. Fast training of neural networks for remote sensing , 1994 .
[56] Tohru Nitta,et al. An Extension of the Back-Propagation Algorithm to Complex Numbers , 1997, Neural Networks.
[57] Kiyoshi Nishiyama,et al. H∞-learning of layered neural networks , 2001, IEEE Trans. Neural Networks.
[58] John S. Shawe-Taylor,et al. Linear programming algorithm for neural networks , 1990, Neural Networks.
[59] Farid U. Dowla,et al. Backpropagation Learning for Multilayer Feed-Forward Neural Networks Using the Conjugate Gradient Method , 1991, Int. J. Neural Syst..
[60] Lijun Liu,et al. Convergence Analysis of Three Classes of Split-Complex Gradient Algorithms for Complex-Valued Recurrent Neural Networks , 2010, Neural Computation.
[61] Mohammad Bagher Menhaj,et al. Training feedforward networks with the Marquardt algorithm , 1994, IEEE Trans. Neural Networks.
[62] Martin Fodslette Møller,et al. A scaled conjugate gradient algorithm for fast supervised learning , 1993, Neural Networks.
[63] Yutaka Fukuoka,et al. A modified back-propagation method to avoid false local minima , 1998, Neural Networks.
[64] Shixin Cheng,et al. Dynamic learning rate optimization of the backpropagation algorithm , 1995, IEEE Trans. Neural Networks.
[65] Mahmood R. Azimi-Sadjadi,et al. Fast learning process of multilayer neural networks using recursive least squares method , 1992, IEEE Trans. Signal Process..
[66] Lawrence D. Jackel,et al. Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.
[67] Jonas Sjöberg,et al. Efficient training of neural nets for nonlinear adaptive filtering using a recursive Levenberg-Marquardt algorithm , 2000, IEEE Trans. Signal Process..
[68] Roberto Battiti,et al. Learning with first, second, and no derivatives: A case study in high energy physics , 1994, Neurocomputing.
[69] Andrew Chi-Sing Leung,et al. Two regularizers for recursive least squared algorithms in feedforward multilayered neural networks , 2001, IEEE Trans. Neural Networks.
[70] M. J. D. Powell,et al. Restart procedures for the conjugate gradient method , 1977, Math. Program..
[71] Yixin Chen,et al. TREAT: a trust-region-based error-aggregated training algorithm for neural networks , 2002, Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290).
[72] L. Dixon. Conjugate Gradient Algorithms: Quadratic Termination without Linear Searches , 1975 .
[73] C. Charalambous,et al. Conjugate gradient algorithm for efficient training of artifi-cial neural networks , 1990 .
[74] Kazumi Saito,et al. Partial BFGS Update and Efficient Step-Length Calculation for Three-Layer Neural Networks , 1997, Neural Computation.
[75] Patrick van der Smagt. Minimisation methods for training feedforward neural networks , 1994, Neural Networks.
[76] Jukka Saarinen,et al. Complex RPROP-algorithm for neural network equalization of GSM data bursts , 2004, Neurocomputing.
[77] Edward W. Kamen,et al. A local linearized least squares algorithm for training feedforward neural networks , 2000, IEEE Trans. Neural Networks Learn. Syst..
[78] Michael T. Manry,et al. An efficient hidden layer training method for the multilayer perceptron , 2006, Neurocomputing.
[79] Okyay Kaynak,et al. An algorithm for fast convergence in training neural networks , 2001, IJCNN'01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222).
[80] E. Polak,et al. Computational methods in optimization : a unified approach , 1972 .
[81] George Cybenko,et al. Ill-Conditioning in Neural Network Training Problems , 1993, SIAM J. Sci. Comput..
[82] Nazif Tepedelenlioglu,et al. A fast new algorithm for training feedforward neural networks , 1992, IEEE Trans. Signal Process..
[83] Thomas P. Vogl,et al. Rescaling of variables in back propagation learning , 1991, Neural Networks.
[84] M Maarten Steinbuch,et al. Uncertainty modelling and structured singular-value computation applied to an electromechanical system , 1992 .
[85] Chin-Sung Liu,et al. Quadratic optimization method for multilayer neural networks with local error-backpropagation , 1999, Int. J. Syst. Sci..
[86] Frank Bärmann,et al. On a class of efficient learning algorithms for neural networks , 1992, Neural Networks.
[87] Theodoula N. Grapsa,et al. Self-scaled conjugate gradient training algorithms , 2009, Neurocomputing.
[88] Marc Toussaint,et al. Rprop Using the Natural Gradient , 2005 .
[89] Henry Leung,et al. The complex backpropagation algorithm , 1991, IEEE Trans. Signal Process..
[90] Mostafa Kaveh,et al. Conjugate gradient learning algorithms for multilayer perceptrons , 1989, Proceedings of the 32nd Midwest Symposium on Circuits and Systems,.
[91] Youmin Zhang,et al. A Fast U-d Factorization-based Learning Algorithm with Applications to Nonlinear System Modeling and Identification , 2022 .
[92] Paul Kang-Hoh Phua,et al. Parallel nonlinear optimization techniques for training neural networks , 2003, IEEE Trans. Neural Networks.
[93] Sammy Siu,et al. Analysis of the Initial Values in Split-Complex Backpropagation Algorithm , 2008, IEEE Transactions on Neural Networks.
[94] Shun-ichi Amari,et al. Natural Gradient Works Efficiently in Learning , 1998, Neural Computation.
[95] Bhaskar D. Rao,et al. A generalized learning paradigm exploiting the structure of feedforward neural networks , 1996, IEEE Trans. Neural Networks.
[96] Ya-Xiang Yuan,et al. A Nonlinear Conjugate Gradient Method with a Strong Global Convergence Property , 1999, SIAM J. Optim..
[97] Lee A. Feldkamp,et al. Decoupled extended Kalman filter training of feedforward layered networks , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[98] Ying Zhang,et al. Boundedness and Convergence of Split-Complex Back-Propagation Algorithm with Momentum and Penalty , 2013, Neural Processing Letters.
[99] Kwok-Wo Wong,et al. A pruning method for the recursive least squared algorithm , 2001, Neural Networks.
[100] Don R. Hush,et al. Error surfaces for multilayer perceptrons , 1992, IEEE Trans. Syst. Man Cybern..
[101] Martin Fieldhouse,et al. Review: Book review: Linear programming , 1964, Comput. J..
[102] David F. Shanno,et al. Conjugate Gradient Methods with Inexact Searches , 1978, Math. Oper. Res..
[103] S. Ergezinger,et al. An accelerated learning algorithm for multilayer perceptrons: optimization layer by layer , 1995, IEEE Trans. Neural Networks.
[104] Tülay Adali,et al. Fully Complex Multi-Layer Perceptron Network for Nonlinear Signal Processing , 2002, J. VLSI Signal Process..