Complex-Valued Neural Networks: Learning Algorithms and Applications

Complex-valued data arise in various applications, such as radar and array signal processing, magnetic resonance imaging, communication systems, and processing data in the frequency domain. To deal with such data properly, neural networks are extended to the complex domain, referred to as complex-valued neural networks (CVNNs), allowing the network parameters to be complex numbers and the computations to follow the complex algebraic rules. Unlike the real-valued case, the nonlinear functions in the CVNNs do not have standard complex derivatives as the Cauchy-Riemann equations do not hold for them. Consequently, the traditional approach for deriving learning algorithms reformulates the problem in the real domain which is often tedious. In this thesis, we first develop a systematic and simpler approach using Wirtinger calculus to derive the learning algorithms in the CVNNs. It is shown that adopting three steps: (i) computing a pair of derivatives in the conjugate coordinate system, (ii) using coordinate transformation between real and conjugate coordinates, and (iii) organizing derivative computations through functional dependency graph greatly simplify the derivations. To illustrate, a gradient descent and LevenbergMarquardt algorithms are considered. Although a single-layered network, referred to as functional link network (FLN), has been widely used in the real domain because of its simplicity and faster processing, no such study exists in the complex domain. In the FLN, the nonlinearity is endowed in the input layer by constructing linearly independent basis functions in addition to the original variables. We design a parsimonious complex-valued FLN (CFLN) using orthogonal least squares (OLS) method, where the basis functions are multivariate polynomial terms. It is observed that the OLS based CFLN yields simple structure with favorable performance comparing to the multilayer CVNNs in several applications. It is well known and interesting that a complex-valued neuron can solve several nonlinearly separable problems, including the XOR, parity-n, and symmetry detection problems, which a real-valued neuron cannot. With this motivation, we perform an empirical study of classification performance of single-layered CVNNs on several real-world benchmark classification problems with two new activation functions. The experimental results exhibit that the classification performances of single-layered CVNNs are comparable to those of multilayer real-valued neural networks. Further enhancement of discrimination ability has been obtained using the ensemble approach. I dedicate this dissertation to my parents

[1]  A.V. Oppenheim,et al.  The importance of phase in signals , 1980, Proceedings of the IEEE.

[2]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[3]  W. Wirtinger Zur formalen Theorie der Funktionen von mehr komplexen Veränderlichen , 1927 .

[4]  Jagdish C. Patra,et al.  A functional link artificial neural network for adaptive channel equalization , 1995, Signal Process..

[5]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.

[6]  Francesco Piazza,et al.  On the complex backpropagation algorithm , 1992, IEEE Trans. Signal Process..

[7]  Kazuyuki Murase,et al.  Real-Time Hand Gesture Recognition Using Complex-Valued Neural Network (CVNN) , 2011, ICONIP.

[8]  Ken Kreutz-Delgado,et al.  The Complex Gradient Operator and the CR-Calculus ECE275A - Lecture Supplement - Fall 2005 , 2009, 0906.4835.

[9]  Xin Yao,et al.  Evolutionary ensembles with negative correlation learning , 2000, IEEE Trans. Evol. Comput..

[10]  Tohru Nitta,et al.  Solving the XOR problem and the detection of symmetry using a single complex-valued neuron , 2003, Neural Networks.

[11]  Colin Giles,et al.  Learning, invariance, and generalization in high-order neural networks. , 1987, Applied optics.

[12]  Howard E. Michel,et al.  Enhanced artificial neural networks using complex numbers , 1999, IJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339).

[13]  Kazuyuki Murase,et al.  Wirtinger Calculus Based Gradient Descent and Levenberg-Marquardt Learning Algorithms in Complex-Valued Neural Networks , 2011, ICONIP.

[14]  B. A. D. H. Brandwood A complex gradient operator and its applica-tion in adaptive array theory , 1983 .

[15]  Hualiang Li,et al.  Algorithms for Complex ML ICA and Their Stability Analysis Using Wirtinger Calculus , 2010, IEEE Transactions on Signal Processing.

[16]  Cris Koutsougeras,et al.  Complex domain backpropagation , 1992 .

[17]  R. Polikar,et al.  Ensemble based systems in decision making , 2006, IEEE Circuits and Systems Magazine.

[18]  Lutz Prechelt,et al.  A Set of Neural Network Benchmark Problems and Benchmarking Rules , 1994 .

[19]  Sundaram Suresh,et al.  A Fully Complex-Valued Radial Basis Function Network and its Learning Algorithm , 2009, Int. J. Neural Syst..

[20]  Kazuyuki Aihara,et al.  Complex-valued prediction of wind profile using augmented complex statistics , 2009 .

[21]  Gavin Brown,et al.  Diversity in neural network ensembles , 2004 .

[22]  V D Calhoun,et al.  Independent component analysis of fMRI data in the complex domain , 2002, Magnetic resonance in medicine.

[23]  José A. Macías,et al.  Evolution of functional link networks , 2001, IEEE Trans. Evol. Comput..

[24]  J. Peixoto Hierarchical Variable Selection in Polynomial Regression Models , 1987 .

[25]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[26]  M. Morris Mano,et al.  Computer Logic Design , 1972 .

[27]  K. Scharnhorst,et al.  Angles in Complex Vector Spaces , 1999 .

[28]  Sung-Bae Cho,et al.  A comprehensive survey on functional link neural networks and an adaptive PSO–BP learning for CFLNN , 2010, Neural Computing and Applications.

[29]  Yoh-Han Pao,et al.  The functional link net and learning optimal control , 1995, Neurocomputing.

[30]  Joos Vandewalle,et al.  Multi-Valued and Universal Binary Neurons: Theory, Learning and Applications , 2012 .

[31]  D. Signorini,et al.  Neural networks , 1995, The Lancet.

[32]  Tohru Nitta,et al.  An Extension of the Back-Propagation Algorithm to Complex Numbers , 1997, Neural Networks.

[33]  Yoh-Han Pao,et al.  Adaptive pattern recognition and neural networks , 1989 .

[34]  Sergios Theodoridis,et al.  Ieee Transactions on Signal Processing Extension of Wirtinger's Calculus to Reproducing Kernel Hilbert Spaces and the Complex Kernel Lms , 2022 .

[35]  Narasimhan Sundararajan,et al.  A self-regulated learning in Fully Complex-valued Radial Basis Function Networks , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[36]  Roberto Battiti,et al.  First- and Second-Order Methods for Learning: Between Steepest Descent and Newton's Method , 1992, Neural Computation.

[37]  D. Opitz,et al.  Popular Ensemble Methods: An Empirical Study , 1999, J. Artif. Intell. Res..

[38]  Akira Hirose,et al.  Complex-Valued Neural Networks (Studies in Computational Intelligence) , 2006 .

[39]  Lajos Hanzo,et al.  Fully complex-valued radial basis function networks: Orthogonal least squares regression and classification , 2008, Neurocomputing.

[40]  Søren Nørvang Madsen,et al.  Radar options for global earthquake monitoring , 2002, IEEE International Geoscience and Remote Sensing Symposium.

[41]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[42]  Narasimhan Sundararajan,et al.  Fully complex extreme learning machine , 2005, Neurocomputing.

[43]  Sung-Bae Cho,et al.  Evolutionarily optimized features in functional link neural network for classification , 2010, Expert Syst. Appl..

[44]  Hualiang Li,et al.  Complex-Valued Adaptive Signal Processing Using Nonlinear Functions , 2008, EURASIP J. Adv. Signal Process..

[45]  Kazuyuki Murase,et al.  Complex-valued functional link network design by orthogonal least squares method for function approximation problems , 2011, The 2011 International Joint Conference on Neural Networks.

[46]  Tülay Adali,et al.  Approximation by Fully Complex Multilayer Perceptrons , 2003, Neural Computation.

[47]  S. Haykin Adaptive Filters , 2007 .

[48]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[49]  Yoh-Han Pao,et al.  Stochastic choice of basis functions in adaptive function approximation and the functional-link net , 1995, IEEE Trans. Neural Networks.

[50]  Complex-Valued Neural Networks: Theories and Applications— , 2006 .

[51]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[52]  Narasimhan Sundararajan,et al.  Communication channel equalization using complex-valued minimal radial basis function neural networks , 2002, IEEE Trans. Neural Networks.

[53]  Akira Hirose,et al.  Complex-Valued Neural Networks: Theories and Applications , 2003 .

[54]  Sheng Chen,et al.  Orthogonal least squares methods and their application to non-linear system identification , 1989 .

[55]  Einar Hille,et al.  Introduction to complex analysis , 1983 .

[56]  O. Mangasarian,et al.  Multisurface method of pattern separation for medical diagnosis applied to breast cytology. , 1990, Proceedings of the National Academy of Sciences of the United States of America.

[57]  T.,et al.  Training Feedforward Networks with the Marquardt Algorithm , 2004 .

[58]  Alan F. Murray,et al.  Enhanced MLP performance and fault tolerance resulting from synaptic weight noise during training , 1994, IEEE Trans. Neural Networks.

[59]  R. Remmert,et al.  Theory of Complex Functions , 1990 .

[60]  Robert A. Jacobs,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.

[61]  A. Bos Complex gradient and Hessian , 1994 .

[62]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[63]  Ganapati Panda,et al.  Nonlinear channel equalization for QAM signal constellation using artificial neural networks , 1999, IEEE Trans. Syst. Man Cybern. Part B.

[64]  Tho Le-Ngoc,et al.  Polynomial perceptrons and their applications to fading channel equalization and co-channel interference suppression , 1994, IEEE Trans. Signal Process..

[65]  Xin Yao,et al.  A constructive algorithm for training cooperative neural network ensembles , 2003, IEEE Trans. Neural Networks.

[66]  Saleem A. Kassam,et al.  Channel Equalization Using Adaptive Complex Radial Basis Function Networks , 1995, IEEE J. Sel. Areas Commun..

[67]  David J. Spiegelhalter,et al.  Machine Learning, Neural and Statistical Classification , 2009 .

[68]  Nils J. Nilsson,et al.  Learning Machines: Foundations of Trainable Pattern-Classifying Systems , 1965 .

[69]  C. Lee Giles,et al.  An analysis of noise in recurrent neural networks: convergence and generalization , 1996, IEEE Trans. Neural Networks.

[70]  Sheng Chen,et al.  Complex-valued radial basis function networks , 1993 .

[71]  Naonori Ueda,et al.  Generalization error of ensemble estimators , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).

[72]  Xin Yao,et al.  Ensemble learning via negative correlation , 1999, Neural Networks.

[73]  Daesik Hong,et al.  Nonlinear blind equalization schemes using complex-valued multilayer feedforward neural networks , 1998, IEEE Trans. Neural Networks.

[74]  Emile Fiesler,et al.  The Interchangeability of Learning Rate and Gain in Backpropagation Neural Networks , 1996, Neural Computation.

[75]  Iku Nemoto,et al.  Complex neural networks , 1992, Systems and Computers in Japan.

[76]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.