Stochastic configuration broad learning system and its approximation capability analysis

In this paper, a kind of stochastic configuration broad learning system (SCBLS) is proposed for data modeling. The proposed SCBLS is established in the form of a flat network and its architecture is determined by a constructive learning approach. The input parameters of feature nodes and enhancement nodes of SCBLS are randomly assigned in the light of a supervisory mechanism. Inequality constraints are used to randomly assign the hidden parameters and adaptively select the scopes of random parameters. The output parameters of SCBLS are determined either by a constructive manner or by solving a global least squares problem. It is proved that the proposed SCBLS possesses universal approximation properties. The performances of the proposed SCBLS are evaluated by function approximation, benchmark datasets and time series prediction. Numerical examples show that SCBLS can achieve satisfactory approximation accuracy.

[1]  Dianhui Wang,et al.  Deep stacked stochastic configuration networks for lifelong learning of non-stationary data streams , 2019, Inf. Sci..

[2]  Fuzhen Zhuang,et al.  Clustering in extreme learning machine feature space , 2014, Neurocomputing.

[3]  C. L. Philip Chen,et al.  Universal Approximation Capability of Broad Learning System and Its Structural Variations , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[4]  Lei Chen,et al.  Enhanced random search based incremental extreme learning machine , 2008, Neurocomputing.

[5]  Ming Li,et al.  Insights into randomized algorithms for neural networks: Practical issues and common pitfalls , 2017, Inf. Sci..

[6]  Xin Liu,et al.  A Linguistic-Valued Approximate Reasoning Approach for Financial Decision Making , 2017, Int. J. Comput. Intell. Syst..

[7]  Shuang Feng,et al.  Fuzzy Broad Learning System: A Novel Neuro-Fuzzy Model for Regression and Classification , 2020, IEEE Transactions on Cybernetics.

[8]  Witold Pedrycz,et al.  A two stage forecasting approach for interval-valued time series , 2018, J. Intell. Fuzzy Syst..

[9]  Fan Min,et al.  Three-way active learning through clustering selection , 2020, International Journal of Machine Learning and Cybernetics.

[10]  C. L. Philip Chen,et al.  Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[11]  Junfei Qiao,et al.  Constructive algorithm for fully connected cascade feedforward neural networks , 2016, Neurocomputing.

[12]  Fuzhen Zhuang,et al.  Parallel extreme learning machine for regression based on MapReduce , 2013, Neurocomputing.

[13]  Witold Pedrycz,et al.  An Integrated neural network with nonlinear output structure for interval-valued data , 2020 .

[14]  Ming Li,et al.  Robust stochastic configuration networks with kernel density estimation for uncertain data regression , 2017, Inf. Sci..

[15]  Zhulin Liu,et al.  Stacked Broad Learning System: From Incremental Flatted Structure to Deep Model , 2021, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[16]  Xiaoping Ma,et al.  Driving amount based stochastic configuration network for industrial process modeling , 2020, Neurocomputing.

[17]  Tie Qiu,et al.  Recurrent Broad Learning Systems for Time Series Prediction , 2020, IEEE Transactions on Cybernetics.

[18]  Paulo C. Rech,et al.  Period-adding and spiral organization of the periodicity in a Hopfield neural network , 2015, Int. J. Mach. Learn. Cybern..

[19]  Xin Wen,et al.  Linguistic truth-valued intuitionistic fuzzy reasoning with applications in human factors engineering , 2016, Inf. Sci..

[20]  C. L. Philip Chen,et al.  Structured Manifold Broad Learning System: A Manifold Perspective for Large-Scale Chaotic Time Series Analysis and Prediction , 2019, IEEE Transactions on Knowledge and Data Engineering.

[21]  Dianhui Wang,et al.  Stochastic Configuration Networks: Fundamentals and Algorithms , 2017, IEEE Transactions on Cybernetics.

[22]  Chee Kheong Siew,et al.  Universal Approximation using Incremental Constructive Feedforward Networks with Random Hidden Nodes , 2006, IEEE Transactions on Neural Networks.

[23]  Li Zou,et al.  Knowledge reasoning approach with linguistic-valued intuitionistic fuzzy credibility , 2020, Int. J. Mach. Learn. Cybern..

[24]  Yoh-Han Pao,et al.  Stochastic choice of basis functions in adaptive function approximation and the functional-link net , 1995, IEEE Trans. Neural Networks.

[25]  Ivan Tyukin,et al.  Approximation with random bases: Pro et Contra , 2015, Inf. Sci..

[26]  Dejan J. Sobajic,et al.  Learning and generalization characteristics of the random vector Functional-link net , 1994, Neurocomputing.

[27]  Chuen-Jyh Chen Structural vibration suppression by using neural classifier with genetic algorithm , 2012, Int. J. Mach. Learn. Cybern..

[28]  Kaoru Ota,et al.  Exponential Stability of Mixed Time-Delay Neural Networks Based on Switching Approaches. , 2020, IEEE transactions on cybernetics.

[29]  Mohamad Khalil,et al.  Parameter selection algorithm with self adaptive growing neural network classifier for diagnosis issues , 2013, Int. J. Mach. Learn. Cybern..