Introduction

Crude oil is a complex mixture which makes it very difficult to achieve thorough characterization at the level of individual molecular type (Ahmadi and Shadizadeh 2012). In order to rectify this issue, group type analysis is deemed as a viable option for crude oil characterization. SARA separation technique is an example of such a group type analysis in which the crude oil is characterized in terms of four distinct classes of saturates, aromatics, resins, and asphaltenes. This separation is implemented based on differences in polarity and solubility (Rasuli Nokandeh et al. 2012; Mohammadi et al. 2012). When dealing with crude oil fractions, most of the attention is focused on asphaltenes mainly due to its adverse influences on the downstream and upstream sectors of petroleum industry. Asphaltenes are known as the heaviest and the most polar portion of crude oil (Kurup et al. 2012) and are defined as fractions of crude oil that are soluble in benzene but insoluble in n-heptane (Shirani et al. 2012a, b). Under initial reservoir conditions, asphaltenes are equilibrated in crude oil through peptizing by resins (Kurup et al. 2012). With changes in thermodynamic conditions such as pressure, temperature, and crude oil composition, stable dispersion of asphaltenes in crude oil would be destabilized; hence, asphaltenes precipitate out of the crude oil solution and settle in the form of solid deposits which can impede the production process (Shirani et al. 2012a, b). In upstream operations, precipitation and deposition of asphaltenes in porous medium during natural depletion and EOR processes substantially affect porosity and permeability of the reservoir rock, and consequently results in loss of efficiency of the recovery process (Kord et al. 2012, 2014; Zanganeh et al. 2012; Moradi et al. 2012). Asphaltene precipitation and deposition also arise in production facilities and transportation pipelines which result in considerable increase in cost of production operation (Abu Tarboush and Husein 2012). To alleviate problems associated with asphaltene precipitation and deposition, two fundamental questions are posed: under what conditions asphaltenes will precipitate out of the crude oil? and what is the extent of asphaltene precipitation and deposition under particular circumstances? Several researchers attempted to identify mechanisms responsible for asphaltenes precipitation and deposition as well as to develop a potent and reliable representative model to quantitatively and qualitatively estimate the extent of asphaltene precipitation and deposition. However, no practical and accurate model is proposed yet to clarify the phase behavior of asphaltene precipitation mainly due to its complex structure and properties (Shirani et al. 2012a, b). The available models for delineation of asphaltenes precipitation are categorized into two distinct classes: models that involve the use of asphaltene properties and the ones that are based on scaling approaches.

Models adopting the asphaltenes properties can be subsumed into two broad categories:

  1. (a)

    Crude oil is a real solution with asphaltenes dissolved in it: The major assumption in these models is that asphaltene precipitation is a reversible process (Mansoori 1997).

  2. (b)

    Asphaltenes are suspended in crude oil because they are peptized by resins: In these models, asphaltene precipitation is considered an irreversible process (Wu et al. 1998).

The second group of the predictive models is the one which is based on scaling equations. In these models, asphaltene precipitation is predicted without being concerned about the complex nature of the asphaltene agglomerates (Rassamdana et al. 1999).

Intelligence systems have been recently used as mathematical predictive tools to quantitatively formulate the amount of asphaltene precipitation using titration data (Ashoori et al. 2010; Naimi et al. 2014; Ahmadi 2011, 2012; Asoodeh et al. 2014a, b; Abedini et al. 2010; Zahedi et al. 2009; Chamkalani et al. 2013; Gholami et al. 2013; Gholami et al. 2014a, b; Fatahi et al. 2014). Although these predictive models are valuable, the quest for greater accuracy has been always remained an issue. In this paper, ACE was firstly used as a potent model to estimate the amount of asphaltene precipitation based on a series of titration experiments using stock tank oils. This non-parametric statistical regression method enhances the fitting by suggesting the optimal transformations for predictor variables (i.e., titration data including temperature, solvent molar weight, and dilution ratio in our case study) and response variable (i.e., the amount of asphaltene precipitation for the case under study in this paper) (Malallah et al. 2006). Secondly, the SVR method was used as a supervised learning method for predicting the amount of asphaltene precipitation. The SVR method is based on the employment of the structural risk minimization principle; hence, it provides better generalization performance in comparison to the conventional Artificial Neural Networks (ANNs) in which the empirical risk minimization principle is implemented. In the end, PLCM method was proposed to combine the outputs associated with SVR and ACE models. To assign appropriate weight factors to each of these two models, GA tool was employed as the combiner. The weight factor associated with each individual model demonstrates the contribution of that model in the final quantitative estimation of asphaltene precipitation. The predicted asphaltene precipitation results using these three models were compared with the experimental data to validate the effectiveness and robustness of the proposed models.

Materials and methods

In this section, application of each of the three predictive methods of ACE, SVR, and PLCM in estimating the amount of asphaltene precipitation is described followed by a description on input/output data space used in this study.

Alternating conditional expectation (ACE) algorithm

This algorithm was developed by Breiman and Friedman as an advanced statistical technique for solving the regression problems containing an unidentified relationship between the predictor and the response variable (Breiman and Friedman 1985). This method has found growing application in different industries owing to its remarkable performance including its capability in appropriate demonstration of a non-linear relationship, if it does exist, between variables in regression problems as well as providing improved fit to the experimental data in comparison to the conventional linear models (Eissa and Shokir 2007). This method achieves the maximum coefficient of multiple correlation through approximating the optimal transformations for the dependent and independent variables (Tang and Zhou 2013).

Generally, a linear regression model for p independent variables of \(X_{1} , X_{2} , \ldots , X_{p}\) and a response variable, Y, is given in the following form (Breiman and Friedman 1985):

$$Y = \beta_{0} + \sum\limits_{i = 1}^{p} {\beta_{i} X_{i} + \varepsilon },$$
(1)

in which β i i = 0 − p are the regression coefficients to be estimated, and ɛ is an error term. In ACE algorithm, the relation between θ(Y) and \(\phi_{1} (X_{1} ), \ldots , \phi_{p} (X_{p} )\) is computed in lieu of assessing the correlation between Y and \(X_{1} , X_{2} , \ldots , X_{p}\) which is conducted in the conventional regression analysis. Based on Eq. (1), the general form of non-parametric ACE algorithm is defined in the following form:

$$\theta (Y) = \alpha + \sum\limits_{i = 1}^{p} {\phi (X_{i} ) + \varepsilon },$$
(2)

in which \(\theta (Y), \phi_{1} (X_{1} ), \ldots , \phi_{p} (X_{p} )\) are the arbitrary measurable mean-zero functions of \(Y,\,X_{1} ,\, X_{2} , \ldots , \,X_{p}\), respectively. The main intention of using ACE algorithm is to seek the optimal transformations, \(\phi_{i}^{*} (X_{i} ), i = 1, \ldots ,\,p\) and θ *(Y), which lead to the maximum correlation between the transformed dependent variable and sum of the transformed predicted variables. This is equivalent to minimizing the value of error variance (ε 2). The value associated with the error variance of a linear regression of the transformed dependent variable on the sum of the transformed independent variables under the constraint of E[θ 2(Y)] = 1 is given by the following equation (Breiman and Friedman 1985):

$$\varepsilon^{2} (\theta , \phi_{i} , \ldots , \phi_{p} ) = {{E\left( {\left[ {\theta (Y) - \sum\limits_{i = 1}^{p} {\phi (X_{i} )} } \right]} \right)^{2} } \mathord{\left/ {\vphantom {{E\left( {\left[ {\theta (Y) - \sum\limits_{i = 1}^{p} {\phi (X_{i} )} } \right]} \right)^{2} } {E\theta^{2} (Y)}}} \right. \kern-0pt} {E\theta^{2} (Y)}}.$$
(3)

ACE algorithm minimizes ε 2 by holding E[θ 2(Y)] = 1, (Y) =  1(X 1) = … =  p (X p ) = 0 through the use of a series of single-function minimizations involving bivariate conditional expectations. Thus, for a given set functions ϕ 1(X 1), …, ϕ p (X p ), minimization of ε 2 with respect to θ(Y) yields the following equation:

$$\theta (Y) = {{E\left[ {\sum\limits_{i = 1}^{p} {\phi (X_{i} )} |Y} \right]} \mathord{\left/ {\vphantom {{E\left[ {\sum\limits_{i = 1}^{p} {\phi (X_{i} )} |Y} \right]} {\left\| {E\left[ {\sum\limits_{i = 1}^{p} {\phi (X_{i} )} |Y} \right]} \right\|}}} \right. \kern-0pt} {\left\| {E\left[ {\sum\limits_{i = 1}^{p} {\phi (X_{i} )} |Y} \right]} \right\|}}.$$
(4)

On the other hand for a given θ(Y), minimization of ε 2 with respect to a single function ϕ j (X j ) yields the following equation:

$$\phi_{j} (X_{j} ) = E\left[ {\theta (Y) - \sum\limits_{i \ne j}^{p} {\phi_{i} (X_{i} )|X_{j} } } \right].$$
(5)

Having implemented the iteration process of minimizing ε 2, the real-valued measurable zero-mean functions of \(\phi_{i} (X_{i} ), i = 1, \ldots , p\) and θ(Y) are determined which are based on determination of the optimal transformation of \(\phi_{i}^{*} (X_{i} ), i = 1, \ldots ,p\) and θ *(Y). In the transformed space, the response and predictor variables are related to each other according to the following equation:

$$\theta^{*} (Y) = \sum\limits_{i = 1}^{p} {\phi_{i}^{*} (X_{i} ) + e^{*} },$$
(6)

in which e * is the error which is not captured by the use of ACE transformations and is assumed to have a normal distribution with zero mean (Breiman and Friedman 1985).

Support vector regression (SVR) algorithm

SVR is a novel learning methodology based upon statistical learning theory which is applicable for solving the complex non-linear regression estimation problems (Asoodeh et al. 2014a, b, c; Bagheripour et al. 2015). This method was first introduced as a robust tool by Vapnik (2000) and has found drastic applicability in the field of regression estimation owing to its many attractive features and higher generalization ability in comparison to traditional ANNs. The difference between generalization performance of SVR and ANN is attributed to the difference in principle of risk minimization implemented in these two models. In SVR, the structural risk minimization is implemented for minimizing the upper bound of generalization error, while ANN takes the advantage of empirical risk minimization to minimize the prediction error (Asoodeh and Bagheripour 2013).

The main role of the SVR algorithm is to estimate a function which computes the functional dependency between an output variable “y,” defined on “R,” and n-dimensional input variable “x.” The decision function f(xw) can be approximated using the following equation (Kecman 2005):

$$f(x,w) = w^{T} x + b,$$
(7)

in which w and b denote the weight vector and the bias term, respectively, and w T denotes the transposition of weight matrices.

For computing f(xw), it is necessary to calculate the value of unknown parameters of w and b. In the SVR algorithm, the following regularized risk function should be minimized in order to determine w and b:

$$R(C) = \frac{1}{2}\left\| w \right\|^{2} + C\sum\limits_{i = 1}^{l} {L_{\varepsilon } (y_{i} ,f_{i} (x,w))},$$
(8)

in which \(C\frac{1}{l}\sum\nolimits_{i = 1}^{l} {L_{\varepsilon } (y_{i} ,f_{i} (x,w))}\) is empirical error, \(\frac{1}{2}\left\| w \right\|^{2}\) is the measure of function flatness, and l is the total number of sample. The constant C > 0 is the penalty parameter which computes the trade-off between the empirical error and the model complexity. ε-insensitive loss function L ɛ (y i f i (x, w)) is used for estimation of empirical error and is determined using the following equation (Kecman 2005):

$$\left| {y - f(x,w)} \right|_{\varepsilon } = {\text{ }}\left\{ {\begin{array}{*{20}l} 0 \hfill & {{\text{if}}\;\left| {y - f(x,w)} \right| \le \varepsilon } \hfill \\ {\left| {y - f(x,w)} \right| - \varepsilon } \hfill & {{\text{Otherwise}}} \hfill \\ \end{array} } \right. .$$
(9)

Minimizing Eq. (8) is equivalent to minimizing the following equation:

$$R = \frac{1}{2}\left\| w \right\|^{2} + C\left( {\sum\limits_{i = 1}^{l} {\xi_{i} + \sum\limits_{i = 1}^{l} {\xi_{i}^{*} } } } \right).$$
(10)
$${\text{Under the constrains of}}\,\left\{ {\begin{array}{*{20}l} {y_{i} - w^{T} x_{i} - b \le \varepsilon + \xi _{i},} & {\quad i = 1, \ldots ,l} \\ {w^{T} x_{i} + b - y_{i} \le \varepsilon + \xi _{i}^{*},} & {\quad i = 1, \ldots ,l} \\ {\xi _{i} \ge 0,\xi _{i}^{*} \ge 0,} & {\quad i = 1, \ldots ,l} \\ \end{array} } \right.,$$

in which ξ and ξ * are the positive slack variables that are employed for training data points above and below the regression tube, respectively. Slack variables are determined as follows:

$$\xi = \left| {y - f(x,w)} \right| - \varepsilon\quad {\text{for data above an }} \,\varepsilon \text{ - }{\text{tube}},$$
(11)
$$\xi^{*} = \left| {y - f(x,w)} \right| - \varepsilon\quad {\text{for data below an }} \varepsilon \text{ - }{\text{tube}}.$$
(12)

For solving Eq. (10), a primal Lagrangian variable is formed first, followed by substituting the Karush–Kuhn–Tucker conditions in primal Lagrangian variable. Once substituted, the problem is formulated in dual space in the form of the following function (Kecman 2005):

$${\text{Maximize}}\,\,L_{p} (\alpha_{i} ,\alpha_{i}^{*} ) = - \frac{1}{2}\sum\limits_{i,j = 1}^{l} {\left( {\alpha_{i} - \alpha_{i}^{ *} } \right)\left( {\alpha_{j} - \alpha_{j}^{ *} } \right)x_{i}^{T} x_{j} } - \varepsilon \sum\limits_{i = 1}^{l} {\left( {\alpha_{i} + \alpha_{i}^{ *} } \right)} + \sum\limits_{i = 1}^{l} {\left( {\alpha_{i} - \alpha_{i}^{ *} } \right)y_{i} }.$$
(13)
$${\text{Subject to}}\,\left\{ {\begin{array}{*{20}l} {\sum\limits_{i = 1}^{l} {(\alpha_{i} - \alpha_{i}^{ *} ) = 0} } \\ {0 \le \alpha_{i} \le C,\quad i = 1, \ldots ,l} \\ { 0 \le \alpha_{i}^{*} \le C,\quad i = 1, \ldots ,l} \\ \end{array} } \right.,$$

in which α i α * i  ≥ 0 are called Lagrangian multipliers.

From the above equation, one Lagrange multiplier pair is calculated. By the use of this Lagrange multiplier pair, the optimal desired weight factor of the regression hyperplane is determined as follows:

$$w_{o} = \sum\limits_{i = 1}^{l} {\left( {\alpha_{i} - \alpha_{i}^{*} } \right)x_{i} }.$$
(14)

The training points which satisfy α i  − α * i  ≠ 0 are called free support vectors. These points lead to determination of the bias term b. Owing to numerical sensitivity of bias term computation, it is preferable to calculate the bias term by averaging over all the free support vectors. Hence, the best hyperplane regression estimation model becomes (Kecman 2005):

$$f\left( {x,w} \right) = w_{o}^{T} x + b = \sum\limits_{i = 1}^{l} {\left( {\alpha_{i} - \alpha_{i}^{ *} } \right)x_{i}^{T} x} + b.$$
(15)

A generalization of the SVR method to non-linear regression is performed through kernel function. Kernel function maps the input vector x ∊ R n into vector ϕ(x) of high dimension feature space F and then solves a linear regression problem in this feature space. Several kernel functions are employed in SVR namely linear, polynomial, radial basis function (RBF), and sigmoid. By employment of the kernel function G and the Lagrangian multiplier pairs of α i α * i , the best non-linear regression function is computed as follows:

$$f\left( {x,w} \right) = G\left( {\alpha - \alpha^{*} } \right) + b.$$
(16)

For determining the bias term b, it is recommended to average over all the free support vectors (Kecman 2005).

Committee machine (CM) algorithm

This method has been recognized as a promising approach in solving estimation problems owing to its better performance in comparison to its individual elements that are executed for prediction target. The essence of the CM algorithm is that it has a parallel framework which produces a final output by combining the results associated with individual models (Asoodeh and Bagheripour 2012a, b; Gholami et al. 2014a, b). Every predictive algorithm such as ANN, empirical correlation, and other modeling algorithms can construct the element of CM (Chen and Lin 2006). By combining the output of individual experts, CM reaps the benefits of each expert. Weight factor of each model indicates its contribution in the final prediction. Optimal contribution of each algorithm in the overall prediction is determined through Genetic Algorithm, GA (Asoodeh, 2013). The schematic diagram of CM is demonstrated in Fig. 1 (Haykin 1999).

Fig. 1
figure 1

A schematic diagram of CM (Haykin 1999)

Genetic algorithm (GA)

The GA procedure was first introduced by John Holland in 1975 and is a robust optimization technique based on principle of natural selection and evolution (Holland 1975). This method has gained prominence as an effective searching technique owing to its high capability of solving non-linear and discontinuities problems in widespread fields of research. The simple GA cycle comprises of three stages of selection, genetic operation, and replacement. Optimization is started with initial population of random chromosomes. Computation of each chromosome fitness value is fulfilled by objective function(s). The best parent solutions are reproduced and the new population is generated through generic operation. Generic operation employs two processes to generate children namely as crossover and mutation. Crossover permits the exchange of information among individuals in the population and provides the innovative capability of GA. Mutation ensures desirable diversity. The children solutions are evaluated and the whole cycle repeats until desirable termination criterion is attained (Kadkhodaie-Ilkhchi et al. 2009; Ghiasi-Freez et al. 2012). A general flowchart of the GA procedure is shown in Fig. 2.

Fig. 2
figure 2

A general flowchart of GA (Chen and Lin 2006)

Input/output data space

The asphaltene precipitation data used to construct the predictive models in this study were borrowed from the open literature (Hu and Guo 2001). The authors added normal alkanes with different carbon numbers to stock tank oil at different dilution ratios and temperatures in the form of a series of titration tests. Asphaltenes precipitated in these titration tests were then separated from the solution using Whatman filter paper No. 42. The composition of oil used in this study is provided in Table 1. The amount of asphaltene precipitation was then modeled using a scaling equation. In this modeling approach, the input variables consist of dilution ratio “R v ,” temperature “T,” and molecular weight of solvent “M w ,” while the amount of asphaltene precipitation “W t ” was regarded as the target value. Each of these input variables had qualitative effects on the amount of the target value according to the sets of data used in this study. For instance, increasing the dilution ratio increased the amount of asphaltene precipitation until the entire asphaltene content of the crude oil sample was precipitated out of the solution. By increasing the molecular weight of the solvent, the amount of asphaltene precipitation decreased. Temperature has a dual impact on the amount of asphaltene precipitation. While many researchers contended that an increase in temperature would increase the amount of asphaltene precipitation, some others believe the reverse trend would be the case. According to Hu and Guo’s study, increasing the temperature decreased the amount of asphaltene precipitation. The data collected from the literature was divided into two classes of training data and testing data points. 70 % of the data points (i.e., 124 data points in total) were employed to construct the model whereas the remainder 30 % (i.e., 52 data points in total) was used to test the model. The statistics of the data used in this study is provided in Table 2.

Table 1 Compositional characteristics and properties of the degassed Caoqiao crude oil and separator gas (Hu and Guo 2001)
Table 2 Statistical description of the dataset used for developing the predictive models

In order to check the validity and accuracy of the asphaltene precipitation data predicted using PLCM model with ACE and SVR as the constituents, two statistical parameters were defined namely as Correlation Coefficient (R 2) and Mean Square Error (MSE). Correlation Coefficient (R 2) illustrates how good the model corresponds the experimental data, hence represents a measure of the utility of the model. The closer the value of R 2 to 1, the better the model fits the data. This parameter is mathematically represented by

$$R^{2} = 1 - \frac{{\sum\limits_{i = 1}^{n} {\left( {Y_{{i {\text{meas}}}} - Y_{{i {\text{pred}}}} } \right)}^{2} }}{{\sum\limits_{i = 1}^{n} {\left( {\overline{Y} - Y_{{i {\text{pred}}}} } \right)}^{2} }},$$
(17)

in which Y meas, Y pred, \(\overline{Y},\) and n are measured variables, estimated variables, mean of measured variables, and number of data points, respectively.

Mean Square Error (MSE) computes the data dispersion around zero deviation and is defined as

$${\text{MSE}} = \frac{1}{n}\sum\limits_{i = 1}^{n} {\left( {Y_{{i {\text{meas}}}} - Y_{{i {\text{pred}}}} } \right)^{2} }.$$
(18)

Results and discussion of results

ACE model predictions

The ACE procedure was first employed to construct the model. To achieve the best correlation between the predictor variables and the response value, the ACE procedure has the capability of using the independent and dependent variables not only in their actual space but also in their logarithmic space. After testing all possible combinations of the predictor and response variables in either logarithmic or actual space, the optimal combination is computed as follows:

$$W_{t} = f\left( {T, R_{v} , M_{w} } \right).$$
(19)

The optimal transformations for temperature, dilution ratio, solvent molecular weight, and amount of asphaltene precipitation, calculated by the use of ACE algorithm, are presented in Fig. 3a–d. These figures helped to extract the formulation between the values of certain parameters and optimal transformation associated with them. The optimal transformation of W t versus summation of the optimal transformations of T, R v , and M w is presented in Fig. 4. The deviation of some data points from the straight line correlation demonstrates the error which could not be captured by ACE transformation method. Furthermore, the comparison cross-plot of the amount of asphaltene precipitation which is predicted using the ACE procedure versus the corresponding set of experimental data is presented in Fig. 5. The acceptable agreement between the predicted and actual measurements is observed.

Fig. 3
figure 3

Optimal transformation of a dilution ratio, b molecular weight of solvent, c temperature, and d amount of asphaltene precipitation computed by ACE

Fig. 4
figure 4

Optimal transformation of W t versus the sum of the optimal transformations of R v, M w , and T

Fig. 5
figure 5

Cross-plot illustrating the correlation coefficient between measured asphaltene precipitation values and those predicted by the ACE procedure

To develop a computational model for predicting the amount of asphaltene precipitation, it seems crucial to seek a correlation between the optimal data transforms and the actual data values. For this purpose, the optimal data transforms were related to the original data through the use of curve fitting. The resulting equations for temperature, dilution ratio, and molecular weight of solvent with respect to their corresponding coefficients are

$$\phi^{*} \left( {R_{v} } \right) = a_{1} {\text{EXP}}\left( {a_{2} \times R_{v} } \right) + a_{3} {\text{EXP}}\left( {a_{4} \times R_{v} } \right),$$
(20)

in which a 1 = 0.28350000, a 2 = 0.05531000, a 3 = −4.858, and a 4 = −0.29660000,

$$\phi^{*} \left( {M_{w} } \right) = a_{3} \left( {M_{w} } \right)^{2} + a_{2} \left( {M_{w} } \right) + a_{1},$$
(21)

in which a 3 = 0.00031400, a 2 = −0.22320000, and a 1 = 39.02000000,

$$\phi^{*} (T) = a_{3} (T)^{2} + a_{2} (T) + a_{1},$$
(22)

in which a 3 = 0.00031400, a 2 = −0.22320000, and a 1 = 39.02000000.

The linear regression between the above independent variables and the dependent variable of the amount of asphaltene precipitation is given as follows

$$\theta^{*} (W_{t} ) = \phi^{*} (T) + \phi^{*} (R_{v} ) + \phi^{*} (M_{w} ).$$
(23)

The numerical value of the amount of asphaltene precipitation is determined by employing the following polynomial correlation:

$$W_{t} = a_{3} \left( {\theta^{*} (W_{t} )} \right)^{2} + a_{2} \left( {\theta^{*} (W_{t} )} \right) + a_{1},$$
(24)

in which a 3 = 0.13230000, a 2 = 1.65900000, and a 1 = 2.81900000.

Table 3 illustrates the R 2 and MSE values for the ACE model for the training and testing stages. Moreover, a comparison between the measured and predicted amounts of asphaltene precipitation versus different sample numbers is demonstrated in Fig. 6. According to the data presented in Figs. 5, 6 and also in Table 3, ACE model has a good performance in terms of predicting the amount of asphaltene precipitation.

Table 3 Comparison between PLCM model and the individual ACE and SVR models in term of the statistical parameters for the training and testing stages
Fig. 6
figure 6

A comparison between the measured amounts of asphaltene precipitation and the predicted values using the ACE model

SVR model predictions

A detailed study about formulating the values of asphaltene precipitation associated with the titration data through the use of SVR was performed by Naimi et al. (2014). The same analogy was used in this section using our selected set of experimental data borrowed from Hu and Guo (2001). An epsilon SVR model was employed in this section to predict the amount of titration-based asphaltene precipitation based on a series of experimental data. Radial basis function (RBF) was chosen as the kernel function owing to its simplicity and greater performance. Optimum range for RBF performance is [−1 1] prior to applying the constructed model. Grid search and pattern search are some surveying techniques for computing the optimal values of parameters involved in RBF namely as C, Gamma, and Epsilon. The search range and extracted optimal values for these involved parameters are reported in Table 4.

Table 4 Properties of optimal SVR model

To evaluate the performance of the SVR model, titration data from Hu and Guo (2001) were used as inputs for the developed model and the amount of asphaltene precipitation was predicted. The cross-correlation between the measured and SVR-predicted amount of asphaltene precipitation is illustrated in Fig. 7. The statistical parameters of R 2 and MSE for the SVR model predictions are also shown in Table 3. Significantly high value of correlation coefficient as well as low values associated with MSE confirms the success of SVR modeling in estimating the amount of asphaltene precipitation. A comparison between the measured and predicted values of asphaltene precipitation for different sample numbers is also shown in Fig. 8. This figure shows that SVR is a potent model to predict the amount of asphaltene precipitation.

Fig. 7
figure 7

Cross-plot illustrating the correlation coefficient between measured asphaltene precipitation values and those predicted by the SVR model

Fig. 8
figure 8

A comparison between the measured amounts of asphaltene precipitation and the predicted values using the SVR model

Power-law committee machine (PLCM) model predictions

A PLCM model was developed in this study by adopting the concept of “divide and conquer” to combine the outputs of ACE and SVR models. Different studies showed that breaking down the data space into subspaces and solving each separately would lead to increasing accuracy of final prediction (Asoodeh and Bagheripour 2012a, b). To construct the proposed PLCM model, the following equation was introduced to GA:

$${\text{MSE}}_{\text{PLCM}} = \frac{1}{k}\sum\limits_{i = 1}^{k} {\left( {w_{1} \times {\text{ACE}}^{{w_{2} }}_{i} + w_{3} \times {\text{SVR}}^{{w_{4} }}_{i} - T_{i} } \right)^{2} },$$
(25)

in which w i (i = 1, …, 4) refer to weight factors assigned to each particular output of the predictive models, k denotes the total number of data, and T i refers to the corresponding target value.

The GA which is implanted in the body of the CM model is able to extract optimum values of the aforementioned weight factors in such a way that the CM model would cover to the lowest MSE associated with the prediction. GA initiates with random population of 50 potential chromosomes (i.e., solutions) and evolves toward the best chromosome through the use of genetic operations. Genetic operators, including elite preservation and crossover were regulated to 2 and 0.76, respectively. In other words, 20 % of the succeeding generation is composed of mutation and inversion operations. The execution of the GA procedure for the proposed PLCM model to predict titration-based asphaltene precipitation is shown in Fig. 9.

Fig. 9
figure 9

The best and mean predicted fitted values associated with the amount of asphaltene precipitation after 1000 generations

After running the GA procedure, the optimal weight factors associated with PLCM model were extracted. Weights which were assigned to each model in Eq. 25 are shown in Table 5. These values demonstrate the optimal contribution of each model in final predictions. By applying these weight factors to the outputs of the individual models of ACE and SVR, the amount of asphaltene precipitation for new titration datasets was predicted and then was compared against the experimental data. A diagram of the proposed PLCM model for the purpose of predicting the amount of asphaltene precipitation is shown in Fig. 10. To assess the reliability of the constructed PLCM model, asphaltene precipitation testing datasets were used as input feed data for the PLCM model. The cross-plot composed of the predicted values of asphaltene precipitation versus the measured amounts from the titration experiments is shown in Fig. 11.

Table 5 Weights assigned to each model’s contribution to optimize the final prediction
Fig. 10
figure 10

Schematic diagram of the PLCM model designed in this study

Fig. 11
figure 11

Cross-plot illustrating the correlation coefficient between measured asphaltene precipitation values and those predicted by the PLCM model

The comparison between the statistical parameters representing the performance of the PLCM model as opposed to the individual models of ACE and SVR is provided in Table 3. The trends presented in Fig. 11 as well as the statistical indicators provided in Table 3 prove that the PLCM predicted results are more reliable compared to the other individual models. In other words, in situations where there are multiple options to solve a problem, it is possible to construct a PLCM model by spending little additional computation effort to the expense of having enhanced accuracy in final predictions. In addition, Fig. 12 demonstrates a comparison between the measured and predicted amounts of asphaltene precipitation versus different data samples. From this Figure, it is concluded that PLCM model is a powerful tool for predicting the amount of asphaltene precipitation in titration experiments.

Fig. 12
figure 12

A comparison between the measured amounts of asphaltene precipitation and the predicted values using the PLCM model

Conclusions

Three smart techniques were used in this study to predict the amount of asphaltene precipitated during titration experiments. It was concluded that PLCM model is capable of predicting the amount of asphaltene precipitation with higher accuracy considering the statistical analysis performed on each prediction. The results predicted using the PLCM model were compared against those obtained using the other two methods of ACE and SVR which proved the superior performance of the PLCM model compared to the other two predicting procedures.