Process modeling and optimization using focused attention neural networks

Abstract Neural networks have been shown to be very useful for modeling and optimization of nonlinear and even chaotic processes. However, in using standard neural network approaches to modeling and optimization of processes in the presence of unmeasured disturbances, a dilemma arises between achieving the accurate predictions needed for modeling and computing the correct gains required for optimization. As shown in this paper, the Focused Attention Neural Network (FANN) provides a solution to this dilemma. Unmeasured disturbances are prevalent in process industry plants and frequently have significant effects on process outputs. In such cases, process outputs often cannot be accurately predicted from the independent process input variables alone. To enhance prediction accuracy, a common neural network modeling practice is to include other dependent process output variables as model inputs. The inclusion of such variables almost invariably benefits prediction accuracy, and is benign if the model is used for prediction alone. However, the process gains , necessary for optimization, sensitivity analysis and other process characterizations, are almost always incorrect in such models. We describe a neural network architecture, the FANN, which obtains accuracy in both predictions and gains in the presence of unmeasured disturbances. The FANN architecture uses dependent process variables to perform feed-forward estimation of unmeasured disturbances, and uses these estimates together with the independent variables as model inputs. Process gains are then calculated correctly as a function of the estimated disturbances and the independent variables. Steady-state optimization solutions thus include compensation for unmeasured disturbances. The effectiveness of the FANN architecture is illustrated using a model of a process with two unmeasured disturbances and using a model of the chaotic Belousov–Zhabotinski chemical reaction.

[1]  Eduardo D. Sontag,et al.  Feedback Stabilization Using Two-Hidden-Layer Nets , 1991, 1991 American Control Conference.

[2]  K S Narendra,et al.  IDENTIFICATION AND CONTROL OF DYNAMIC SYSTEMS USING NEURAL NETWORKS , 1990 .

[3]  Paul J. Werbos,et al.  Neurocontrol and fuzzy logic: Connections and designs , 1992, Int. J. Approx. Reason..

[4]  Richard Weber,et al.  The use of secondary measurements to improve control , 1972 .

[5]  David E. Rumelhart,et al.  Predicting the Future: a Connectionist Approach , 1990, Int. J. Neural Syst..

[6]  J. Roux,et al.  Experimental studies of bifurcations leading to chaos in the Belousof-Zhabotinsky reaction , 1983 .

[7]  Kumpati S. Narendra,et al.  Identification and control of dynamical systems using neural networks , 1990, IEEE Trans. Neural Networks.

[8]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[9]  James D. Keeler,et al.  Predicting the Future: Advantages of Semilocal Units , 1991, Neural Computation.

[10]  L. H. Ungar,et al.  PROCESS MODELING AND CONTROL USING NEURAL NETWORKS , 1996 .

[11]  James D. Keeler,et al.  Layered Neural Networks with Gaussian Hidden Units as Universal Approximations , 1990, Neural Computation.

[12]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[13]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[14]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .