1 Introduction

The multimodel approach is an interesting alternative and a powerful tool for representation, observation, emulation and control of complex processes.

The basic idea of this approach is the decomposition of the full operation range of the process into a number of operating regimes. In each operating regime, a simple local model is considered. These local models called model’s library are then combined in some way to yield a global model.

Nonlinear models are also widely used in engineering science applications to describe the dynamic behavior of real processes. Due to their mathematical complexity, they are not easily exploitable for designing an observer, an emulator, a control law or for setting up a diagnosis strategy, even if they give a good description of the considered process. Assuming that the modeled process evolves around an operating point, a linearization procedure is then possible and leads to the reduction of the mathematical complexity of the nonlinear model. Hence, analysis tools available for linear systems can be used. However, in practical cases, the linearity assumption is not always valid and consequently the linearized model does not represent the whole behavior of the process[1, 2].

Moreover, it will be more interesting in practice to have a model that is easily usable and gives a good global characterization of the process dynamic behavior.

Different approaches exist in the literature dealing with the multimodel representation. Initially, the founder of the multimodel approach focused on the regime decomposition and the choice of local model structure on the base of a combination of qualitative, quantitative, empirical and mechanistic system knowledge[35]. A fuzzy set methods for local modeling and identification of nonlinear systems were described in [1]. In which, local models structure was described by means of fuzzy “if-then” rules and any prior knowledge about the operating regime was not required. Other approaches based on the Takagi-Sugeno model were developed in [6, 7]. For the representation of discrete and continuous linear systems with uncertain parameters, some methods related to the systematic generation of a models’ base were proposed in [813].

Two basic multimodel structures for interconnecting the sub-models (local models) can be distinguished[14].

  1. 1)

    In the first structure, the sub-models have the same state vector (Takagi-Sugeno multimodels). This leads to the same structure for all sub-models. Which limits the choice of the models’ base. Several works dealing with modeling, state estimation, control and diagnosis of nonlinear systems were based on this multimodel structure[7, 1518].

  2. 2)

    In the second one, the sub-models are uncoupled and their state vectors are different (uncoupled multimodel). So we can consider sub-models with different structures. As works dealing with this multimodel structure, we cite [1923].

In the two aforementioned multimodel structures, the contribution of the sub-models depends on the weighting functions. The modeling performances are closely related to the choice of the centers and degrees of overlap between them (for example, the dispersions for Gaussian functions case). In this frame work, we propose a new method to optimize these parameters for the systematic generation of weighting function centers and dispersions. This method integrates a classification procedure of an identification data set picked out of the considered system.

The paper is organized as follows: We present in Section 2, the principle of the multimodel representation. The proposed method for an optimal systematic generation of the weighting function centers and dispersions is detailed in Section 3. The next section presents the results of a numerical simulation example. An experimental validation on a semi-batch reactor is given in Section 5. Conclusions will be drawn in Section 6.

2 Multimodel representation

“Divide and conquer” strategy is the basic idea of the multimodel approach. It consists in subdividing the full operating range of the system into several operating zones. At each operating zone a simple linear sub-model is identified using the known tools of linear systems identification, linearization techniques or other techniques. The global model, which is valid under all operating conditions of the system, is thereafter formed by a fusion of the sub-models and it is easily exploited to study any local proprieties and to deduce simple and powerful observers, emulators and controls. These sub-models are typical, but not always selected. The multimodel output (the output of the global model) is generated through an interpolation mechanism that integrates an online weighting functions evaluation.

2.1 Weighting functions

The sub-models contributions can be quantified by the weighting functions μ i (ξ(k)) that ensure smooth transition between sub-models. It can be obtain, for example, from the normalized Gaussian function:

$${\mu _i}(\xi (k)) = {{{\omega _i}(\xi (k))} \over {\sum\limits_{j = 1}^N {{\omega _j}(\xi (k))} }}$$
(1)

where

$${\omega _i}(\xi (k)) = \exp \left({ - {{{{(\xi (k) - {c_i})}^2}} \over {\sigma _i^2}}} \right)$$

where c i is the centre of the i-th weighting function and σ i is the dispersion for weighting functions, N is the number of models. The weighting functions satisfy the following constraint[1, 3, 4, 1922, 24]:

$$\matrix{{\sum\limits_{i = 1}^N {{\mu _i}(\xi (k))} = 1,\quad \forall k} \cr {0 \leqslant {\mu _i}(\xi (k)) \leqslant 1,\quad \forall k,\quad \forall i = 1, \cdots, N} \cr }$$

where ξ(k) is the decision variable. It can be an exogenous measured signals: for example, input, output and state variable. A large choice of weighting functions is possible such as Gaussian functions, sigmoidal functions, etc. There are various ways of connecting sub-models in order to generate the multimodel output y MM (k). We distinguish two multimodel structures according to the use of coupled or uncoupled states.

2.2 State multimodel structure

Two multimodel structures are possible. The coupled state multimodel structure also called Takagi-Sugeno multimodel[3, 4, 24] is extensively used in the multimodel analysis and synthesis. This multimodel structure is given by

$$\matrix{{x(k + 1) = \sum\limits_{i = 1}^N {{\mu _i}(\xi (k))} \left\{ {{A_i}x(k) + {B_i}u(k)} \right\}} \cr {{y_{MM}}(k) = \sum\limits_{i = 1}^N {{\mu _i}(\xi (k))} {C_i}x(k)} \cr }$$
(2)

where xR n is the common state vector of the sub-models, uR m the input, y MM R p is the multimodel output, A i R n ×n, B i R n ×m and C i R p ×n are constant matrices. In this multimodel framework, the contribution of each sub-model is evaluated through a weighted sum of submodels parameters. Therefore, this multimodel structure can be considered as a system with time varying parameters (we note that there is only a global state x MM ) and we have

$$\matrix{{x(k + 1) = \tilde A(k)x(k) + \tilde B(k)u(k)} \cr {{y_{MM}}(k){\rm{ }} = \tilde C(k)x(k){\rm{ }}} \cr }$$
(3)

where \(\tilde A(k) = \sum _{i = 1}^N{\mu _i}(\xi (k)){A_i},\,\tilde B(k) = \sum _{i = 1}^N{\mu _i}(\xi (k)){B_i}\) and \(\tilde C(k) = \sum _{i = 1}^N{\mu _i}(\xi (k)){C_i}\).

The other possible structure is called the uncoupled state multimodel[14, 1922]:

$$\matrix{{{x_i}(k{\rm{ }} + {\rm{ }}1){\rm{ }} = {A_i}(k){x_i}(k) + {B_i}(k)u(k)} \cr {{y_i}(k){\rm{ }} = {C_i}(k){x_i}(k)} \cr }$$
(4)

and

$${y_{MM}}(k) = \sum\limits_{i = 1}^N {{\mu _i}(\xi (k))} {y_i}(k)$$
(5)

where \({x_i} \in {{\bf{R}}^{{n_i}}}\) and y i R p are respectively the state vector and the output vector of the i-th sub-model, uR m is the input, y MM R p is the multimodel output, \({A_i} \in {{\bf{R}}^{n \times n}},\,{B_i} \in {{\bf{R}}^{n \times m}}\) and \({C_i} \in {{\bf{R}}^{p \times n}}\) are constant matrices. In this case the multimodel output is formulated by a weighted sum of sub-models outputs. Therefore, each sub-model evolves independently in its own state space depending on the input control and its initial state. The main interest of the uncoupled state multimodel structure is the decoupling between the sub-models. Indeed, different submodels structure can be used.

2.3 Parametric estimation

Essentially, there are two estimation criteria can be used in the parameter identification of the uncoupled multi-model: global and local criteria[25]. The retained estimation criterion, is the global one defined by

$$J = {1 \over 2}\sum\limits_{k = 1}^{{N_H}} {{{({y_{MM}}\left(k \right) - y\left(k \right))}^2}} $$
(6)

where N H is the number of training data and y(k) is the real output of the nonlinear system.

This criterion favors a good global characterization between global behavior of the nonlinear system and the multimodel. In this section, we present the parameter estimation procedure based on the uncoupled state multimodel approach. The column vector θ is the vector of unknown parameters of the multimodel.

$$\theta = {[\theta _1^{\rm{T}}\, \cdots \,\theta _i^{\rm{T}}\, \cdots \,\,\theta _N^{\rm{T}}]^{\rm{T}}}$$

where each column block θ i is formed by the q i unknown parameters of the particular sub-model denoting the unknown scalar parameter of the sub-model i.

$${\theta _i} = {[{\theta _{i,1}} \cdots {\theta _{i,q}} \cdots {\theta _{i,{q_i}}}]^{\rm{T}}}$$
(7)

where θ i,q denotes the unknown scalar parameter of the sub-model i.

The parametric estimation of θ is based on an iterative minimization procedure of a quadratic criterion J(θ) which can be a Levenberg-Marquardt’s algorithm[26] considering the persistent excitation condition.

$$\theta (it + 1) = \theta (it) - \Delta (it){(H(\theta) + \lambda (it)I)^{ - 1}}G(\theta)$$
(8)

where θ(it) is the vector of parameters at a particular iteration it, H(θ) is Hessian matrix, and G(θ) is the gradient vector. The calculus of the gradient vector and the Hessian matrix are based on the calculation of sensitivity functions with respect to sub-models parameters.

The gradient G is given by the simple derivation of the global criterion with respect to parameters θ as

$$G = {{\partial J} \over {\partial \theta }} = \sum\limits_{k = 1}^{{N_H}} {\varepsilon (k)} {{\partial {y_{MM}}(k)} \over {\partial \theta }}$$
(9)

where

$$\varepsilon (k) = ({y_{MM}}(k) - y(k)).$$
(10)

Hessian matrix H is obtained by a double derivation of the global criterion with respect to parameters θ as

$$H = \sum\limits_{k = 1}^{{N_H}} {{{\partial {y_{MM}}(k)} \over {\partial \theta }}} {{\partial {y_{MM}}(k)} \over {\partial {\theta ^{\rm{T}}}}}$$
(11)

where

$${{\partial {y_{MM}}(k)} \over {\partial \theta }} = \sum\limits_{i = 1}^N {{\mu _i}(\xi (k))} {{\partial {y_i}(k)} \over {\partial \theta }}.$$
(12)

3 Optimal systematic generation of weighting function centers

The multimodel representation is composed of a set of sub-models. Each one is valid in a specified operating zone. The local validity of a sub-model is defined by a weighting function which tends to one in the corresponding operating zone and tends to zero outside it. The weighting function related to each sub-model can be defined in classical cases by the static characteristic or by linearization procedure. Also, they can be defined systematically using clustering techniques and fuzzy clustering techniques. The difficulties related to these last solutions are the determination of the number of clusters and their centers which will be the weighting functions centers. In this work, we propose a method for the optimal determination of the weighting functions centers based on Chiu’s classification method[9, 27, 28]. Here, the normalized Gaussian functions is retained as

$${\omega _i}(\xi (k)) = \exp \left({ - {{{{(\xi (k) - ci)}^2}} \over {{\sigma _i}^2}}} \right)$$
(13)

and

$${\mu _i}(\xi (k)) = {{{\omega _i}(\xi (k))} \over {\sum\limits_{j = 1}^N {{\omega _j}(\xi (k))} }}$$
(14)

where c i is the center of the i-th weighting function, σ i is the corresponding dispersion, and ξ is the decision variable. The choice of centers c i has a great influence on the modeling precision.

3.1 Influence of centers c i choice

To demonstrate the influence of the centers, we choose them based on the system static characteristic. Indeed, after picking up the nonlinear system static characteristic, we proceed to subdivide it into linear zones. Each linear zone which will be represented by a sub-model is associated with a weighting function centered on the number c i . This center is also chosen based on the system static characteristic.

3.1.1 Identification example

Let us use the uncoupled multimodel structure to represent the behavior of the nonlinear system[29]:

$$y\left({k + 1} \right) = {{y\left(k \right)} \over {1 + {y^2}\left(k \right)}} + {u^3}\left(k \right),\quad u \in \left[ { - 1\,\,\,\,1} \right]$$
(15)

Here, the identification of the multimodel is realized by a global criterion. The weighting functions μ i depend on the input signal ξ(k) = u(k), and the centers are c 1 = − 0.3, c 2 = 0.0, c 3 = 0.3 and the dispersion σ = 0.1. Fig. 1 shows the resulting weighting functions evolutions. A set of input/output data points is used to build the multimodel. Fig. 2 shows the evolution of this data set). In this case, the signal used to validate the uncoupled state multimodel is given by

$$u(k) = 07\sin ({{2\pi k} \over {100}}).$$
(16)
Fig. 1
figure 1

Weighting functions evolutions (classical choice based on the static characteristic)

Fig. 2
figure 2

Data set for system identification

The resulting local models, by application of the identification algorithm, are described by the following expressions:

Sub-model 1.

$${A_1} = 0.8909,{\rm{ }}{B_1} = 0.1308,{\rm{ }}{C_1} = 1.$$

Sub-model 2.

Sub-model 3.

$${A_3}{\rm{ }} = {\rm{ }}0.8919,{B_3}{\rm{ }} = 0.1357,{C_3}{\rm{ }} = {\rm{ }}1.$$

The validation results are given in Fig. 3. This figure shows the evolutions of the real and the multimodel outputs (y(k) and y MM (k)). The simulation results show that the multimodel output cannot describe the real behavior with satisfactory precision. The mean square error (MSE) and the variance-accounted-for (VAF) have been calculated to evaluate the performance by the following equations:

$${\rm{MSE}} = {1 \over {{N_H}}}\sum\limits_{k = 1}^{{N_H}} {{{\left({{y_{MM}}\left(k \right) - y\left(k \right)} \right)}^2}} $$
(17)
$${\rm{VAF}} = \max \left\{ {1 - {{{\rm{var}}\left({{y_{MM}}\left(k \right) - y\left(k \right)} \right)} \over {{\rm{var}}\left({{y_{MM}}\left(k \right)} \right)}},0} \right\} \times 100\%.$$
(18)
Fig. 3
figure 3

Evolutions of the real and the multimodel outputs

They are respectively equal, in this case, to 0.021 and 92.17%.

3.2 Proposed optimal systematic generation of centers c i

We propose in this work to generate systematically the synthesis parameters (centers and dispersions) related to the weighting functions. Indeed, the full operating range of the nonlinear system is partitioned into a set of operating zones via a classification procedure which provides the cluster centers that will be considered as the weighting function centers. The dispersion of each weighting function is then given in function of the distances separating the neighbor centers.

3.2.1 Classification procedure

Having a set of numerical data, the classification procedure consists in selecting between these data, the cluster centers using Chiu’s method for the numerical data classification[9, 27, 28]. In this work, we aimed to extend this method to classify a set of regression vectors \({\varphi _j}({\varphi _j} = {\left[ {y(j){\rm{ }}y(j - 1){\rm{ }}u(j - 1)} \right]^{\rm{T}}},\,j = 1,\, \cdots, \, M)\) built from a set of persistent identification data covering the hole operating range of the considered nonlinear system.

Each regression vector φ j , is associated with a potential P j given by the expression as

$${P_j} = \sum\limits_{h = 1}^M {{{\rm{e}}^{\left({{{ - 4{{\left\Vert {{\varphi _j} - {\varphi _h}} \right\Vert }^2}} \over {r_a^2}}} \right)}}} $$
(19)

where r a is a positive parameter controlling the decrease ratio of the potential. The potential decreases exponentially as φ h moves away from φ j . The first regression vector cluster center φ 1* is the vector whose potential P 1* given by (19) is the maximum.

To avoid selecting the first cluster center φ 1* having the potential P 1 * and its neighbors as other cluster centers, the classification procedure assigns to each potential P j the following new value:

$${P_j} \leftarrow {P_j} - P_1^*{{\rm{e}}^{\left({{{4{\rm{||}}{\varphi _j} - \varphi _1^*|{|^2}} \over {r_b^2}}} \right)}}.$$
(20)

The parameter r b (r b > 0) must be selected larger than r a to favor the operation related to the selection of other cluster center completely different from the last one. Next, we select, as second cluster-center, the vector whose modified potential given by the relation (20) is the maximum. Let the second cluster-center be \(\varphi_{2}^{*}\) and the associated potential value be \({P}_{2}^{*}\).

In a similar manner, we choose the i-th cluster center \(\varphi_{i}^{*}\) with the maximum potential \({P}_{i}^{*}\) and modify the potentials as follows:

$${P_j} \leftarrow {P_j} - P_i^*{{\rm{e}}^{\left({{{ - 4{\rm{||}}{\varphi _j} - \varphi _i^*|{|^2}} \over {r_b^2}}} \right)}}.$$
(21)

Chiu determined the cluster selection by introducing two positive parameters ε 1 and ε 2 (ε 1 > ε 2). He proposed the following inequalities for stopping the cluster centers’ selection.

  1. 1)

    If \(P_i^* > {\varepsilon _1}P_1^* \Rightarrow \) Selection sanctioned.

  2. 2)

    If \(P_i^* < {\varepsilon _2}P_1^* \Rightarrow \) Selection stopped.

  3. 3)

    If \({\varepsilon _2}P_1^* \leqslant P_i^* \leqslant {\varepsilon _1}P_1^*\) and if:

    $${{{\rm{min}}\left({{\rm{min}}(|\varphi _i^* - \varphi _1^*{\rm{|}}), \cdots, {\rm{min}}(|\varphi _i^* - \varphi _{i - 1}^*{\rm{|}})} \right)} \over {{r_a}}} < 1 - {{P_i^*} \over {P_1^*}}$$

    where \(\varphi_{i}^{*}\) is the current regression vector center and \(\varphi_{1}^{*}\), \(\varphi _1^*,\,\varphi _2^*,\, \cdots, \, \varphi _1^* - 1\) are the last selected ones. The regression vector center to be retained corresponds, in this case, to the maximum value of the potentials after rejecting the current value \({P}_{i}^{*}\).

3.2.2 Centers determination

Once the regression vector cluster centers are selected and referring to the decision variable to be considered (ξ(k) = y(k) or ξ(k) = u(k)), we have the weighting functions centers c i .

$${c_i} = u_i^*{\rm{ or }}{c_i} = y_i^*{\rm{, }}i = 1, \cdots, N.$$

3.2.3 Dispersions computation

Then, we calculate the corresponding dispersion σ i using one expression of the two given below:

Expression 1. The dispersions σ i can be adjusted according to the mean distance to the n nearest neighbors[24].

$${\sigma _i} = \alpha {1 \over n}\sum\limits_{j = 1}^n {\left| {{c_i} - {c_j}} \right|}$$
(22)

where c i is the current center and c j is the nearest neighbor center to c i . α is a scaling factor defining the degree of overlap between the weighting functions.

Expression 2. The dispersions σ i is considered to be in proportion to the distance separating the center c i and the nearest center[28].

$${\sigma _i} = {1 \over \alpha }\mathop {\min }\limits_{j = 1, \cdots, n} \left({\left| {{c_i} - {c_j}} \right|} \right)$$
(23)

where α is a positive factor defining the degree of overlap between the weighting functions. Order estimation of retained models can be insured by instrumental determinants’ ratio-test method[2].

4 Simulation results

We consider the same nonlinear system given by (15). The same identification data set (see Fig. 1) is used to generate the set of regression vectors \({\varphi _j}\,({\varphi _j} = {\left[ {y(j)\,y(j - 1)\,u(j - 1)} \right]^{\rm{T}}})\) for the extended classification procedure to generate the regression vectors cluster centers \(\varphi_{i}^{*}\). It generates three clusters centered on the following regression vectors:

$$\varphi _1^* = \left[ {\matrix{-.8149 \cr - 0.8181\cr - 0.6874 \cr } } \right],\,\,\varphi _2^* = \left[ {\matrix{- 0.0555\cr - 0.0557\cr - 0.0376 \cr } } \right],\,\,\varphi _3^* = \left[ {\matrix{0.4655\cr 0.4646 \cr 0.4369\cr } } \right].$$

For the decision variable ξ(k), we consider the input u(k) (the third element of the regression vector). So, we have the weighting functions centers c i as

$${c_1} = u_1^* = - 0.6874,{c_2} = u_2^* = - 0.0376,{c_3} = u_3^* = 0.4369$$

Using (23), we deduce the corresponding dispersions σ i as

$${\sigma _1} = 0.6498,\quad {\sigma _2} = 0.4745,\quad {\sigma _3} = 0.4745$$

Fig. 4 gives the resulting weighting functions evolution.

Fig. 4
figure 4

Weighting functions evolutions

The multimodel identification is realized with a global criterion (as described previously in the identification procedure) and it adopts the uncoupled state multimodel structure. The resulting sub-models are described in the state space by the following expressions:

Sub-model 1.

$${A_1} = - 0.1468,\quad {B_1} = 1.6142,\quad {C_1} = 1.$$

Sub-model 2.

Sub-model 3.

$${A_3} = 0.5334,\quad {B_3} = 0.6489,\quad {C_3} = 1.$$

The same validation conditions are conserved. Fig. 5 illustrates the multimodel validation results. It gives the evolutions of the multimodel response and the real one. We can observe that the multimodel can describe properly the non linear system behavior. The MSE and the variance-accounted-for VAF have been calculated to evaluate the performance of the proposed method. They are respectively equal to 0.003 and 98.81 %.

Fig. 5
figure 5

Evolutions of the multimodel output y MM (k) and the real output y(k). (Systematic and optimal choice of centers and dispersions)

The obtained results using proposed procedure are very good relatively to the case where the classical procedure is used. The following table (Table 1) summarizes the MSE and VAF calculated in the two cases.

Table 1 MSE and VAF

5 Experimental validation

The satisfactory modeling performances obtained in numerical simulations, incited us to perform an experimental validation on a semi-batch reactor[22, 30, 31]. The plant consists of a perfectly stirred tank provided with a jacket where the heat is exchanged between a cooling fluid and the reaction mixture as shown in Fig. 6. The heating-cooling system fluid flow rate through the jacket is constant and the fluid temperature into the jacket is modulated using an external servo system involving a plate heat exchanger together with electric resistors. The plate exchanger is used for the fluid cooling, namely the tap water, while the heating of the fluid is ensured by the electric resistors. Several temperature sensors allow to measure the reactor temperatures and the inlet and outlet jacked temperatures. The reactor is mainly used in bach mode for chemical esterification of the crude olive oil according to the following reaction:

$$Acid + Alcohol\rightleftharpoons Ester + Water.$$
Fig. 6
figure 6

The chemical reactor

The typical temperature profile for the involved esterification experiment is the following: The reaction mixture is first heated from the room temperature to a specified temperature which usually corresponds to the reaction temperature, then this temperature is maintained constant until the end of the reaction. Finally, the reactor is cooled back to the room temperature.

The control variable is the electric power supplied by the heating resistors while the output is the reactor temperature. Such a system is nonlinear as mentioned in [13, 22, 32]. This is the rational behind the multimodel approach to be adopted.

Fig. 7 gives implementation scheme of models’ base optimal systematic generation for the considered chemical reactor. The set of identification data picked out of the reactor given by Fig. 8 (the process sampling time is 180 s) is treated off line by the classification procedure to generate the weighting functions centers. So, we obtain the following regression vectors centers:

$$\varphi _1^* = \left[ {\matrix{83 \cr 83 \cr 1000 \cr } } \right],\quad \varphi _2^* = \left[ {\matrix{89 \cr 88.1 \cr 1700 \cr } } \right],\quad \varphi _3^* = \left[ {\matrix{104.5 \cr 104.4 \cr 2000 \cr } } \right].$$
Fig. 7
figure 7

Implementation scheme of the proposed method for the chemical reactor

Fig. 8
figure 8

Identification data set

When we choose the input u(k) as decision variable, we can deduce the weighting functions centers.

$${c_1} = u_1^* = 1000,\quad {c_2} = u_2^* = 1700,\quad {c_3} = u_3^* = 2000.$$

We deduce the corresponding dispersions using relation (22).

$${\sigma _1} = 0.0700,\quad {\sigma _2} = 0.0263,\quad {\sigma _3} = 0.0750.$$

Fig. 9 gives the evolution of the resulting weighting functions. The identification procedure application yields to the three sub-models described in the state space as follows:

Fig. 9
figure 9

Evolutions of weighting functions

Sub-model 1.

Sub-model 2

Sub-model 3.

The result of the validation phase is given by Fig. 10 which illustrates the evolution of the multimodel output and the reactor one. We can note that the multimodel output describes the real behavior with satisfactory precision. The variance-accounted-for VAF has been calculated to evaluate the performance of the proposed method. It is equal, in this real case, to 98.7%. Fig. 11 shows the relative errors er MM (%) (between the real output y(k) and the multimodel output y MM (k)). In the full operating regimes (heating, reaction and cooling), the considered error does not exceed 10 %.

Fig. 10
figure 10

Evolutions of the multimodel output y MM (k) and the reactor output y(k)

Fig. 11
figure 11

Evolution of the relative error

6 Conclusions

In this paper, we proposed an optimal systematic generation of the weighting function parameters and the models’ base for an uncoupled state multimodel structure. The proposed method integrates a classification procedure of an identification data set picked out of the considered nonlinear system. The obtained results in numerical simulation are very satisfactory relatively to the case where the system static characteristic is exploited. An experimental validation on a chemical reactor is also judged very satisfactory.