Generalization ability of Boolean functions implemented in feedforward neural networks
暂无分享,去创建一个
[1] Simon Haykin,et al. Neural Networks: A Comprehensive Foundation , 1998 .
[2] Saburo Muroga,et al. Threshold logic and its applications , 1971 .
[3] David Haussler,et al. What Size Net Gives Valid Generalization? , 1989, Neural Computation.
[4] Jacob Feldman,et al. Minimization of Boolean complexity in human concept learning , 2000, Nature.
[5] Alberto L. Sangiovanni-Vincentelli,et al. LSAT-an algorithm for the synthesis of two level threshold gate networks , 1991, 1991 IEEE International Conference on Computer-Aided Design Digest of Technical Papers.
[6] John E. Hopcroft,et al. Synthesis of Minimal Threshold Logic Networks , 1965, IEEE Trans. Electron. Comput..
[7] Robert A. Legenstein,et al. Foundations for a Circuit Complexity Theory of Sensory Processing , 2000, NIPS.
[8] Ian Parberry,et al. Circuit complexity and neural networks , 1994 .
[9] J. Stephen Judd,et al. Neural network design and the complexity of learning , 1990, Neural network modeling and connectionism.
[10] Alberto L. Sangiovanni-Vincentelli,et al. Learning Complex Boolean Functions: Algorithms and Applications , 1993, NIPS.
[11] Noam Nisan,et al. Constant depth circuits, Fourier transform, and learnability , 1993, JACM.
[12] Peter Grassberger,et al. Information and Complexity Measures in Dynamical Systems , 1991 .
[13] Lech Jozwiak,et al. New approach to learning noisy Boolean functions , 1998 .
[14] Lech Jozwiak,et al. An Effective and Efficient Method for Functional Decomposition of Boolean Functions Based on Information Relationships Measures. , 2000 .
[15] Thomas Kailath,et al. Depth-Size Tradeoffs for Neural Computation , 1991, IEEE Trans. Computers.
[16] J. Rice. Mathematical Statistics and Data Analysis , 1988 .
[17] David H. Wolpert,et al. The Mathematics of Generalization: The Proceedings of the SFI/CNLS Workshop on Formal Approaches to Supervised Learning , 1994 .
[18] Wolfgang Kinzel,et al. Antipredictable Sequences: Harder to Predict Than Random Sequences , 1998, Neural Computation.
[19] Michael L. Dertouzos,et al. Threshold Logic: A Synthesis Approach , 1965 .
[20] Nathan Linial,et al. The influence of variables on Boolean functions , 1988, [Proceedings 1988] 29th Annual Symposium on Foundations of Computer Science.
[21] Thomas M. Cover,et al. Elements of Information Theory , 2005 .
[22] László Lovász,et al. INFORMATION AND COMPLEXITY ( HOW TO MEASURE THEM ? ) , 2022 .
[23] Chris Thornton,et al. Parity: The Problem that Won't Go Away , 1996, Canadian Conference on AI.
[24] Peter J. W. Rayner,et al. Generalization and PAC learning: some new results for the class of generalized single-layer networks , 1995, IEEE Trans. Neural Networks.
[25] Ingo Wegener,et al. The complexity of Boolean functions , 1987 .
[26] Robert O. Winder,et al. Threshold logic , 1971, IEEE Spectrum.
[27] Leonardo Franco,et al. Generalization properties of modular networks: implementing the parity function , 2001, IEEE Trans. Neural Networks.
[28] Ronald L. Rivest,et al. Training a 3-node neural network is NP-complete , 1988, COLT '88.
[29] José M. Bravo,et al. Role of Function Complexity and Network Size in the Generalization Ability of Feedforward Networks , 2005, IWANN.
[30] Mathukumalli Vidyasagar,et al. A Theory of Learning and Generalization , 1997 .
[31] Mathukumalli Vidyasagar,et al. A Theory of Learning and Generalization: With Applications to Neural Networks and Control Systems , 1997 .
[32] Mariusz Rawski,et al. Functional decomposition with an efficient input support selection for sub-functions based on information relationship measures , 2001, J. Syst. Archit..
[33] Jirí Síma,et al. Back-propagation is not Efficient , 1996, Neural Networks.
[34] Anders Krogh,et al. Introduction to the theory of neural computation , 1994, The advanced book program.
[35] S. Bhide,et al. A real-time solution for the traveling salesman problem using a Boolean neural network , 1993, IEEE International Conference on Neural Networks.
[36] Ravi B. Boppana,et al. The Average Sensitivity of Bounded-Depth Circuits , 1997, Inf. Process. Lett..
[37] Pekka Orponen,et al. General-Purpose Computation with Neural Networks: A Survey of Complexity Theoretic Results , 2003, Neural Computation.
[38] Sergio A. Cannas,et al. Non glassy ground-state in a long-range antiferromagnetic frustrated model in the hypercubic cell , 2004 .
[39] Pekka Orponen,et al. Computational complexity of neural networks: a survey , 1994 .
[40] C. Lee Giles,et al. What Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagation , 1998 .
[41] Nick Chater. Cognitive science: The logic of human learning , 2000, Nature.
[42] Claude E. Shannon,et al. The synthesis of two-terminal switching circuits , 1949, Bell Syst. Tech. J..
[43] Vladimir Vapnik,et al. Statistical learning theory , 1998 .
[44] Chuanyi Ji,et al. Capacity of Two-Layer Feedforward Neural Networks with Binary Weights , 1998, IEEE Trans. Inf. Theory.
[45] Leonardo Franco,et al. Generalization and Selection of Examples in Feedforward Neural Networks , 2000, Neural Computation.
[46] Noam Nisan,et al. Constant depth circuits, Fourier transform, and learnability , 1989, 30th Annual Symposium on Foundations of Computer Science.
[47] Vinay Deolalikar,et al. Mapping Boolean functions with neural networks having binary weights and zero thresholds , 2001, IEEE Trans. Neural Networks.
[48] Rich Caruana,et al. Overfitting in Neural Nets: Backpropagation, Conjugate Gradient, and Early Stopping , 2000, NIPS.
[49] M. Anthony. Boolean Functions and Artificial Neural Networks , 2003 .