# Multiple neural control and stabilization

- 1.4k Downloads

## Abstract

In this paper, we present a multiple neural control and stabilization strategy for nonlinear and unstable systems. This control strategy method is efficient especially when the system presents different behaviors or different equilibrium points and when we hope to drive the whole process to a desired state ensuring stabilization. The considered control strategy has been applied on a nonlinear unstable system possessing two equilibrium points. It has been shown that the use of the multiple neural control and stabilization strategy increases further the stability domain of the system further than when we use a single neural control strategy.

## Keywords

Behaviors State variables Multiple neural controllers Stabilization## 1 Introduction

Multiple models characterizing different plant operation modes are used to predict the system behaviors. They are the one that best describes the plant and used to initialize new adaptation and/or generate new control input. From a practical point of view, the need to use multiple models in control is often necessary specially when sudden changes in the plant occur, in order to give better performance such as more accurate tracking and larger operation domain. In addition to the abilities of neural networks to imitate nonlinear plant characteristics, both multiple models and neural networks tools have attracted researchers to investigate in the domain of control of complex and nonlinear system especially the field of multiple neural control strategies.

In the 1960s and 1970s, most of works on the multiple model control were based on optimal control. Specifically, problem solving was based on the use of Kalman filters and linear control minimizing a quadratic loss function [1].

In the context of identification, there is no new means used in the multiple model approach, but in the control context, the switching problem was raised and the first proposals were published by Martensson [2]. Following, two types of switching will begin to appear in the literature. The first is known as direct switching, where the choice of the next controller is predetermined and depends on the outputs of the system. The second is known as indirect switching, where local models are used at each moment in which controller will be used [3]. The latter kind of switch is also called supervised control.

There has been a major research activity to extend the multiple model approach in the control field. Narendra et al. [4] presented a general methodology to design a multiple model adaptive control of uncertain systems. This methodology makes systems to operate effectively in an environment with a high degree of uncertainty. As applications, they considered a system described by various behaviors; each behavior is represented by a model including the dynamic relating to the considered environment.

In their book, Murray-Smith and Johansen [5] have made a collection of a number of articles on multiple model approaches. This book considers the various aspects and applications of multiple model approach for modeling, identification and control of nonlinear systems, and it summarizes the theory and application of multiple model adaptive control. In this book, the authors also open up horizons of research on the topic of adaptive multiple model control, such as the determination of the local models number of the complex system to be controlled, the choice of the types of controllers and the validity of the models (weighting functions).

Since 2000, research on multiple model approach has geared toward stability analysis and robust control of systems described by multiple models [6].

This paper is organized such that Sect. 2 presents a general description of a multiple neural control. Section 3 presents the structure of the multiple neural control and stabilization strategy. In Sect. 4, we show a single neural control stabilization method, and in Sect. 5, simulation results are carried out using the multiple neural control and stabilization on an unstable nonlinear system. Finally, a conclusion and prospects are given in Sect. 6.

## 2 Multiple neural control

The application of multiple neural control strategy is based on neural models, which incorporates a set of pair model/controller. Combination and switching between models are the only characteristics of the multiple neural control [4, 5, 7, 8].

The weighting function *f* _{ i }(*x*) (*i* ∈ {1,…, *n*}) represents the validity of the model number *i* (and/or the corresponding controller). In the case when we select a single controller at a given instant (e.g., the *i* ^{ème} controller), the value of the function *f* _{ i }(*x*) is equal to 1 and 0 for all others *f* _{ j }(*x*) (*j* ∈ {1,…, *n*} and *j*≠*i*). The value of the function *f* _{ i }(*x*) is belonging to the range [0, 1] in the case when we combine all models and controllers.

The banks of neural models and neural controllers are made after learning steps from sub-databases representing different behaviors of the controlled system.

## 3 Multiple neural control and stabilization strategy

The multiple neural controller and stabilization strategy is used in order to increase the system’s stabilization domain. It consists on the build of neural controllers of which the learning step is carried out through sub-databases representing different regions (behaviors) of the system.

### 3.1 Principle

The multiple neural control and stabilization strategy is based on the accomplishment of neural controllers direct neural models related to each behavior. The achievement of the learning for controllers and direct models is made from sub-databases identified around the center of each behavior. Each region may be represented in the space of a sphere or an ellipsoid whose center is the desired stabilization point.

### 3.2 Structure of the multiple control and stabilization

*C*(1) is based on the computation of the minimum distance between the current and the desired states, so the selected neural controller

*i*is such that:

*i*and the desired state \(x_{j}^{d}\).

### 3.3 Controller principle

*k*\(u(k) \in R^{m}\) is the control vector and

*f*[.] is the vector of nonlinear functions.

This implies that the states at time *k* + 2 are determined from the states at time *k* and the control values between times *k* and *k* + 2.

*k*+

*N*are determined by the state vector at time

*k*and controls values between times

*k*and

*k*+

*N*− 1.

*U*(

*k*) can be solved according to

*x*(

*k*+

*N*) and

*x*(

*k*). In fact, the authors proved that in this case, the condition of invertibility is interpreted as a local condition freely achievable. In addition, in Ref. [11], it is shown that if the linearized model of the system is controllable and observable, then the local inverse model of the system exists.

*G*is a nonlinear function and Eq. (8) is a fundamental relationship representing the inverse dynamic of the system [10]. The nonlinear function

*G*can be approximated by a neural network. This last will be exploited as a neural controller providing the control actions to stabilize the system around an equilibrium point or a desired state.

### 3.4 Structure of the neural controller

A large class of nonlinear dynamic systems is presented with the state-space models (2), so to realize neural control strategy it is necessary to build neural controllers based on input and states data.

The connection weight *w* _{ j,i } network and *w* _{ o,j } are adjusted in order to minimize a quadratic error criterion (12) between network outputs *Û*(*k*) and the desired outputs *U*(*k*) [10, 11, 12, 13].

The activation function used is sigmoidal one given by (13).

The neural controller output layer contains *M* = *N*·*m* nodes if the dimension of the control vector *u*(*k*) is equal to *m*.

*N* is the necessary number of iterations to evolve the system from the actual state *x*(*k*) to the future state *x*(*k* + *N*).

*Û*(

*k*) to the system.

*W*is the vector connection weights of the neural network.

The number of nodes in the input layer is defined according to the number of current and future states.

The number of nodes of the output layer is equal to the components number of the sequence of control actions applied to the system to reach the future state from the current state.

The number of nodes in the hidden layer is chosen after learning experiences. It is fixed by the structure which gives the lowest value of the error criterion (12).

### 3.5 Learning procedure

Here, sequences of synthesized signal *U*(*k*) are applied as inputs to the system. The corresponding states of these last are used as inputs to the neural network of which the outputs are compared with the training signal (the input of the system). The resulting error is used to adjust the neural network connection weights. This procedure tends to force the neural network to emulate the inverse dynamics of the system.

This learning structure is a classic method of supervised learning, where the teacher (i.e., the synthesis signal) provides directly target values to the output of the learner (i.e., the network model).

*ε*(

*k*) obtained for each learning pattern.

### 3.6 Weights connection adaptation

*w*

_{ ji }. The input value of the

*j*th node in the hidden layer is computed such that:

*w*

_{ ji }is the weight connection between

*j*th node and the

*i*th node of the input layer and

*θ*

_{ j }is the bias value of the node

*j*.

*j*is given by:

*g*is a sigmoidal function (13)

*o*th node in the output layer is computed such that

*w*

_{ oj }between the hidden layer and the output layer is performed as follows:

*w*

_{ ji }between the input layer and the hidden layer is such that:

*η*is the learning rate.

## 4 Single neural stabilization strategy

The controller provides the system an *N* control values in each interval of *N* sampling periods.

The parameter *N* must be at least equal to *n* (order system) for local controllability [11].

This structure of stabilization and control will be used in the case of multiple neural control and stabilization.

## 5 Simulation results

In this section, we present the simulation results after the application of multiple neuronal and stabilization strategy for a nonlinear unstable system extract from Ref. [19].

### 5.1 Presentation of the system

This system presents two equilibrium points \(x_{e1}^{T} = \left( {\begin{array}{*{20}c} 0 & 0 & 0 \\ \end{array} } \right)\) and \(x_{e2}^{T} = \left( {\begin{array}{*{20}c} 1 & 2 & 1 \\ \end{array} } \right)\).

In order to apply the multiple neural control and stabilization strategy, we have made two sub-databases around each equilibrium point (behavior). These sub-databases will be used to learn neural controllers and direct neural local models useful to compute the selection criterion *C* and then to select the appropriate controller.

### 5.2 Sub-databases

The considered system (25) is highly unstable and divergent. The sub-databases are built by varying the system input at a random manner. The inputs should be bounded with low values, so that the system doesn’t diverge from the first iterations. For both sub-databases, we adopt an input signal \(u(k) \in [ - 0,5;\;0,5]\) and different initial states *x*(*k*) chosen near a considered equilibrium point. The future states values *x*(*k* + *N*) are the computed, and all values from time *k* to time *k* + *N* (in our case *N* = 3) are then recorded in text file establishing a sub-database which will be used to learn neural controller and direct neural local model.

For the first sub-database related to the first equilibrium point \(x_{e1}^{T} = \left( {\begin{array}{*{20}c} 0 & 0 & 0 \\ \end{array} } \right)\), the initial states \(x(k) = \left( {\begin{array}{*{20}c} {x_{1} (k)} & {x_{2} (k)} & {x_{3} (k)} \\ \end{array} } \right)^{T}\) are chosen at random values belonging to the interval \([ - 1,1]\) and the second sub-database on the second equilibrium point \(x_{e2}^{T} = \left( {\begin{array}{*{20}c} 1 & 2 & 1 \\ \end{array} } \right)\) where the initial states are chosen at random values belonging, respectively, to the intervals \([0,2]\), \([1,3]\), \([0,2]\).

### 5.3 System response with multiple neural control and stabilization

*n*= 2), minimal distance between the current and the desired states.

Here, the first controller is selected and the system stabilizes around the desired state (first equilibrium point) \(x_{1}^{d}\). In this case, the distance between initial and desired states is \(d_{1} = 0.911\). With the second controller where the distance between initial and second desired states is \(d_{2} = 2.287\), the system diverges.

Note that with the multiple neural control strategy, the system is stabilized around the first equilibrium point (desired state), whereas with a single neural control strategy the system diverges.

We remark that with the single control strategy the states diverge.

Here, the selected controller is the second one, the distance between initial and desired states is \(d_{2} = 1.3\), and the system stabilizes around the desired state (second equilibrium point) \(x_{2}^{d}\). While with the first controller where the distance between initial and desired states is \(d_{1} = 3.59\), the system diverges.

These results confirm the efficient of the multiple neural control and stabilization strategy.

## 6 Conclusion

We have presented a multiple neuronal control and stabilization strategy which can be applied on systems characterized by different behaviors in different regions of the functional domain. In the simulation results, we have considered an unstable system possessing two equilibrium points. So two neural controllers and two direct neural local models have been built around each point; then, the multiple neural control and stabilization strategy has been applied using a distance criterion between the current and the desired states to select the appropriate controller. The use of the latter strategy increases the system stabilization domain further. As future work, we hope to use an adaptive multiple neural control and stabilization strategy with weighting functions representing the validity of models.

## References

- 1.Athans MD, Castanou D, Dunn KP, Greene CS, Lee WH, Sandell NR, Willsky AS (1977) The stochastic control of the F-8C aircraft using a multiple model adaptive control (MMAC) method—Part 1: equilibrium flight. IEEE Trans Autom Control 22(5):768–780CrossRefGoogle Scholar
- 2.Martensson B (1986) Adaptive stabilization. Ph.D. Dissertation. Lund Institute of Technology, LundGoogle Scholar
- 3.Middleton RH, Goodwin GC, Hill DJ, Mayne DQ (1988) Design issues in adaptive control. IEEE Trans Autom Control 33:50–58MathSciNetCrossRefzbMATHGoogle Scholar
- 4.Narendra KS, Balakrishnan J, Ciliz MK (1995) Adaptation and learning using multiple models, switching, and tuning. IEEE Control Syst Mag 15:37–51CrossRefGoogle Scholar
- 5.Murray-Smith R, Johansen TA (1997) Multiple model approaches to modelling and control. Taylor & Francis, OxfordGoogle Scholar
- 6.Kalkkuhl J, Hunt KJ, Zbikowski R, Dzielinski A (1997) Applications of neural adaptive control technology. Series in Robotics and intelligent systems, vol 17. World Scientific, SingaporeGoogle Scholar
- 7.Zhai JY, Shu MF, Xiao HM (2008) Multiple models switching control based on recurrent neural networks. Neural Comput Appl 17:365–371CrossRefGoogle Scholar
- 8.Fourati F (2014) Multiple neural control of a greenhouse. Neurocomputing 139:138–144CrossRefGoogle Scholar
- 9.Li X, Wang S, Wang W (2002) Improved multiple model adaptive control and its stability analysis, 15th Triennal World Congress. S.S.T-Mo-A03, A.C.3b, BarcelonaGoogle Scholar
- 10.Li W, Slotine JJE (1989) Neural network control of unknown nonlinear systems. In: American control conference, vol 2. Pittsburg, pp. 1136–1141Google Scholar
- 11.Levin AU, Narendra KS (1996) Control of nonlinear dynamical systems using neural networks—Part II: observability, identification and control. IEEE Trans Neural Netw 7:30–42CrossRefGoogle Scholar
- 12.Psaltis D, Sideris A, Ymamura AA (1988) A multilayered neural network controller. IEEE Control Syst Mag 8:17–21CrossRefGoogle Scholar
- 13.Nguyen DH, Widrow B (1991) Neural networks for self-learning control systems. IEEE Control Syst Mag 10:18–23CrossRefGoogle Scholar
- 14.Fourati F, Chtourou M, Kamoun M (2007) Stabilization of unknown nonlinear systems using neural networks. Appl Soft Comput 8:1121–1130CrossRefGoogle Scholar
- 15.Hunt KJ, Sbarbaro D, Zbikowski R, Gawthrop PJ (1992) Neural networks for control systems—a survey. Automatica 28:1083–1112MathSciNetCrossRefzbMATHGoogle Scholar
- 16.Xu C, Maoxin L, Qiming Z (2015) On the mean square exponential stability for a stochastic fuzzy cellular neural network with distributed delays and time-varying delays. Int J Innovative Comput Inf Control 11:247–256Google Scholar
- 17.Rumelhart DE, Hinton GE, Williams RJ (1986) Learning internal representations by error propagation. In: Rumelhart DE, McClelland JL, PDP Research Group Corporate (eds) Parallel distributed processing: explorations in the microstructure of cognition, vol 1. MIT Press, Cambridge, pp 318–362Google Scholar
- 18.Yabuta T, Yamada T (1992) Learning control using neural networks. J Syst Eng 2:180–191Google Scholar
- 19.Levin AU, Narendra KS (1993) Control of nonlinear dynamical systems using neural networks: controllability and stabilization. IEEE Trans Neural Netw 4:192–206CrossRefGoogle Scholar

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.