Information dynamics and neural techniques for data analysis

[1]  Gustavo Deco,et al.  An information theory based learning paradigm for linear feature extraction , 1996, Neurocomputing.

[2]  Lucas C. Parra,et al.  Statistical Independence and Novelty Detection with Information Preserving Nonlinear Maps , 1996, Neural Computation.

[3]  Deco,et al.  Statistical-ensemble theory of redundancy reduction and the duality between unsupervised and supervised neural learning. , 1995, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[4]  Gustavo Deco,et al.  Linear redundancy reduction learning , 1995, Neural Networks.

[5]  Gustavo Deco,et al.  Nonlinear higher-order statistical decorrelation by volume-conserving neural architectures , 1995, Neural Networks.

[6]  Deco,et al.  Learning time series evolution by unsupervised extraction of correlations. , 1995, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[7]  M. Paluš Testing for nonlinearity using redundancies: quantitative and qualitative aspects , 1994, comp-gas/9406002.

[8]  J. Theiler,et al.  Generalized redundancies for time series analysis , 1994, comp-gas/9405006.

[9]  L. Parra,et al.  Redundancy reduction with information-preserving nonlinear maps , 1995 .

[10]  Pierre Comon,et al.  Independent component analysis, A new concept? , 1994, Signal Process..

[11]  Ian F. Blake,et al.  Detection in multivariate non-Gaussian noise , 1994, IEEE Trans. Commun..

[12]  W. Ebeling,et al.  Finite sample effects in sequence analysis , 1994 .

[13]  B. Pompe,et al.  On some entropy methods in data analysis , 1994 .

[14]  Ron Meir,et al.  Data compression and prediction in neural networks , 1993 .

[15]  B. Pompe Measuring statistical dependences in a time series , 1993 .

[16]  M. Paluš,et al.  Information theoretic test for nonlinearity in time series , 1993 .

[17]  Ralph Linsker,et al.  Local Synaptic Learning Rules Suffice to Maximize Mutual Information in a Linear Network , 1992, Neural Computation.

[18]  James Theiler,et al.  Testing for nonlinearity in time series: the method of surrogate data , 1992 .

[19]  K. Pawelzik,et al.  Optimal Embeddings of Chaotic Attractors from Topological Considerations , 1991 .

[20]  H. B. Barlow,et al.  Unsupervised Learning , 1989, Neural Computation.

[21]  H. Schuster,et al.  Proper choice of the time delay for the analysis of chaotic time series , 1989 .

[22]  Naftali Tishby,et al.  Consistent inference of probabilities in layered networks: predictions and generalizations , 1989, International 1989 Joint Conference on Neural Networks.

[23]  Ralph Linsker,et al.  How to Generate Ordered Maps by Maximizing the Mutual Information between Input and Output Signals , 1989, Neural Computation.

[24]  H. B. Barlow,et al.  Finding Minimum Entropy Codes , 1989, Neural Computation.

[25]  Martin Casdagli,et al.  Nonlinear prediction of chaotic time series , 1989 .

[26]  P. Szépfalusy,et al.  Characterization of Chaos and Complexity by Properties of Dynamical Entropies , 1989 .

[27]  I. Tsuda,et al.  Calculation of information flow rate from mutual information , 1988 .

[28]  Ralph Linsker,et al.  Self-organization in a perceptual network , 1988, Computer.

[29]  I. Tsuda,et al.  Extended information in one-dimensional maps , 1987 .

[30]  The variance of information loss as a characteristic quantity of dynamical chaos , 1987 .

[31]  J. Rissanen Stochastic Complexity and Modeling , 1986 .

[32]  B. Pompe,et al.  Transinformation of Chaotic Systems , 1986 .

[33]  B. Pompe,et al.  State Predictability and Information Flow in Simple Chaotic Systems , 1986 .

[34]  Fraser,et al.  Independent coordinates for strange attractors from mutual information. , 1986, Physical review. A, General physics.

[35]  Ichiro Tsuda,et al.  Information theoretical approach to noisy dynamics , 1985 .

[36]  D. Ruelle,et al.  Ergodic theory of chaos and strange attractors , 1985 .

[37]  P. Grassberger,et al.  Characterization of Strange Attractors , 1983 .

[38]  Robert Shaw Strange Attractors, Chaotic Behavior, and Information Flow , 1981 .

[39]  F. Takens Detecting strange attractors in turbulence , 1981 .

[40]  J. Rissanen,et al.  Modeling By Shortest Data Description* , 1978, Autom..

[41]  L. Glass,et al.  Oscillation and chaos in physiological control systems. , 1977, Science.

[42]  M. Hénon,et al.  A two-dimensional mapping with a strange attractor , 1976 .

[43]  F. Attneave Some informational aspects of visual perception. , 1954, Psychological review.

[44]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[45]  Claude E. Shannon,et al.  The mathematical theory of communication , 1950 .