Abstract
In gas drilling operations, the rate of penetration (ROP) parameter has an important influence on drilling costs. Prediction of ROP can optimize the drilling operational parameters and reduce its overall cost. To predict ROP with satisfactory precision, a stacked generalization ensemble model is developed in this paper. Drilling data were collected from a shale gas survey well in Xinjiang, northwestern China. First, Pearson correlation analysis is used for feature selection. Then, a Savitzky-Golay smoothing filter is used to reduce noise in the dataset. In the next stage, we propose a stacked generalization ensemble model that combines six machine learning models: support vector regression (SVR), extremely randomized trees (ET), random forest (RF), gradient boosting machine (GB), light gradient boosting machine (LightGBM) and extreme gradient boosting (XGB). The stacked model generates meta-data from the five models (SVR, ET, RF, GB, LightGBM) to compute ROP predictions using an XGB model. Then, the leave-one-out method is used to verify modeling performance. The performance of the stacked model is better than each single model, with R2 = 0.9568 and root mean square error = 0.4853 m/h achieved on the testing dataset. Hence, the proposed approach will be useful in optimizing gas drilling. Finally, the particle swarm optimization (PSO) algorithm is used to optimize the relevant ROP parameters.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Drilling optimization is important, as it can reduce the overall costs of drilling. Increasing the rate of penetration (ROP) is one optimization method. The ROP is affected by various interconnected factors (operational parameters, drill bit characteristics, and formation properties). The prediction of ROP is difficult, and many scholars have spent a lot of effort researching ROP models.
In the early years, some physical and mathematical models were established to predict ROP (Maurer 1962; Bingham 1965; Bourgoyne and Young 1974; Warren 1987; Hareland and Rampersad 1994; Motahhari 2008; Motahhari et al. 2010). The Bourgoyne and Young model considers a variety of influencing factors—mechanical parameters, hydraulic parameters, and formation parameters—and has been widely used in the prediction of ROP (Bahari and Baradaran Seyed 2007; Bahari et al. 2009; Hua 2010; Rahimzadeh et al. 2011; Nascimento et al. 2015; Ahmed and Ibrahim 2019). Soares et al. (2016) compared Hareland and Rampersad’s model (Hareland and Rampersad 1994) and Motahhari’s model (Motahhari 2008; Motahhari et al. 2010) in three different sandstone formations, concluding that the former works best. Al-Abduljabbar et al. (2019) established a robust ROP model using 7000 real-time data measurements from a carbonate formation.
The factors influencing ROP are complex, and physical models have difficulty in integrating them all to accurately predict the ROP. In recent years, machine learning techniques have developed rapidly and are widely used to predict ROP. Hegde et al. (2017) showed that data-driven models were more accurate than physical models (Bingham 1965; Hareland and Rampersad 1994; Motahhari et al. 2010) in ROP prediction in every formation. These data-driven models include linear regression, random forest (RF) (Breiman 2001), and ensemble models (Hegde 2016).
Artificial neural networks (ANNs) are widely used in ROP prediction and have good performance (Arabjamaloei and Shadizadeh 2011; Arabjamaloei et al. 2011; Arabjamaloei and Karimi Dehkordi 2012; Kahraman 2016; Bezminabadi et al. 2017; Anemangely et al. 2018; Elkatatny 2018; Abbas et al. 2019b, a; Sabah et al. 2019; Ashrafi et al. 2019; Diaz et al. 2019; Elkatatny et al. 2020; Zhao et al. 2020; Qian et al. 2021). To improve the performance of ANNs, some algorithms are used to optimize ANN models. Basarir et al. (2014) used an adaptive neuro-fuzzy inference system to predict ROP. Shi et al. (2016) used a typical extreme learning machine and an efficient learning model called upper-layer-solution-aware to predict ROP. Eskandarian et al. (2017) combined RF and monotone multi-layer perceptron models to predict ROP. Anemangely et al. (2018) used a hybrid model composed of a multi-layer perceptron neural network (MLP) together with either a particle swarm optimization (PSO) algorithm or a cuckoo optimization algorithm to predict ROP. Ashrafi et al. (2019) developed and trained eight hybrid ANNs using four evolutionary algorithms: a genetic algorithm, PSO, a biogeography-based optimizer and the imperialist competitive algorithm. Gan et al. (2019a) proposed a novel two-level method that contains a formation drillability fusion submodel established by using the Nadaboost extreme learning machine algorithm and an ROP model established by a neural network with a radial basis function. Elkatatny (2019) developed a new ROP model using an ANN combined with the self-adaptive differential evaluation technique.
Several other machine learning methods, such as support vector regression (SVR) (Vapnik 1995) and ensemble methods such as RF (Breiman 2001) and gradient boosting machines (GB) (Friedman 2001) have also been applied to the prediction of ROP. Bodaghi et al. (2015) proposed an SVR model of ROP that was optimized by a cuckoo search algorithm and genetic algorithm. Ansari et al. (2017) proposed a committee support vector regression improved by an imperialist competitive algorithm. Gan et al. (2019b) proposed a support vector regression with iterative local search and a stochastic inertia weight bat algorithm method. Ahmed et al. (2019) compared four computational intelligent techniques: ANN, ELM, SVR and least-squares SVR. The results show that all four computational intelligent techniques had acceptable accuracy, with the best model being the least-squares SVR. Hegde et al. (2015) used trees, bagging and RF to predict the ROP during drilling. Mantha and Samuel (2016) presented an algorithm that can choose the best of four models—neural networks, SVR, RF or GBM—for use in different strata. They can be effectively employed independently of location or formation. Hegde and Gray (2017) increased ROP by changing surface parameters on a rig that used RF algorithms. Hegde and Gray (2018) built ROP, torque on bit and mechanical specific energy models using a data-driven approach with an RF algorithm. Soares and Gray (2019) studied the real-time predictive capabilities of analytics and machine learning ROP models in a continuous learning environment. It was found that by shortening the retraining interval (defined by the length or number of data points), the performance of analysis models and ML models can be improved.
The drilling model above is suitable for the well-studied in their paper. However, a single-ROP model may not be effective for other wells, as it may fall into a local optimal value. Therefore, to improve the accuracy and generalization of the ROP predictive model, this paper uses a stacked generalization ensemble model. This approach combines six models: SVR, extremely randomized trees (ET), RF, light gradient boosting machine (LightGBM), GB and extreme gradient boosting (XGB). The stacked model includes a two-layer structure. The first layer generates meta-data from the SVR, ET, RF, LightGBM and GB models, and the second layer uses the XGB model to make the final prediction. Then, the PSO algorithm is used to optimize the drilling parameters that are effective influences on the ROP.
Data collection
The data collected from a shale gas survey well drilled in Xinjiang, northwestern China. Table 1 shows the drilling rig model and main auxiliary equipment. The purpose of this well was to determine the shale’s stratigraphic sequence, thickness, structure and organic geochemical characteristics, and to evaluate its storage performance, rock mechanical parameters and gas-bearing properties.
Feature selection
The database consisted of 2383 data points covering 2369 m of drilling depths (10–2379 m). It included the parameters of ROP, depth, weight of bit (WOB), stand pip pressure (SPP), rotary speed (RPM), mud weight (MW), temperature (T), flow rate (Q), bit diameter, equivalent circulating densities (ECD), funnel viscosity (FV), solid content, filter loss (FL), formation pressure gradient, and porosity. As redundant features in the data will affect the performance of the model, some features with lower correlations need to be excluded. Figure 1 shows the values of the Pearson correlations between feature variables. The Pearson correlations of formation pressure gradient and FV with ROP are very low, so these two features were excluded. An overview of the drilling data is shown in Table 2.
Noise reduction
Data measured is always affected by noise. Even in the best-controlled drilling conditions, errors in measurements are at least approximately 5% (Orr 1998; Redman 1998). Noisy data can affect the machine learning process, increase learning time and reduce performance (Garcia et al. 2015). In this research, we used the SG smoothing filter (Savitzky and Golay 1964). The SG filter can smooth a signal without degrading its original properties too much. It is based on the least-squares principle of the polynomial smoothing algorithm. The SG technique is widely used in drilling data preprocessing (Anemangely et al. 2018; Ashrafi et al. 2019; Sabah et al. 2019). According to the characteristics of the data, we filtered and smoothed the following data: ROP, WOB, SPP, MW, temperature, Q, ECD and porosity. Figure 2 compares the data measured in the well (blue) with the data that was denoised using the SG filter (red). It can be seen from the figure that the SG filter has removed the abnormal points and smoothed the curve.
Data normalization
Data normalization is important in machine learning. The numerical magnitudes of each drilling variable are different; therefore, if the data is not normalized, the training time will be longer and the performance of the data-driven model will be poor. It is necessary to normalize raw drilling data to eliminate this effect on the results of the machine learning algorithm. After the raw data have been normalized, each indicator has the same order of magnitude, making it suitable for comprehensive and comparative evaluations. In this paper, we used min–max normalization to normalize the raw drilling data (Sun et al. 2019), as shown in (1).
where xmax is the maximum value and xmin is the minimum value. After normalization, all data are in the interval [0, 1].
Method
In this section, six machine learning models are introduced according to their particular architecture: SVR, RF, ET, GB, LightGBM and XGB. Moreover, the proposed technique—the stacked generalization ensemble model—is comprehensively investigated. Then, the PSO algorithm is introduced to optimize the drilling parameters.
Support vector regression
The support vector machine (SVM) approach was developed by Vanpik and collaborators at Bell Laboratories (Vapnik and Chervonenkis 1964; Vapnik 1995). The SVM principle is based on statistical learning theory and structural minimization (Bello et al. 2016). The SVR model can be defined as the following (Fletcher 2013):
where αi and αi* are Lagrange multipliers and K(xi, x) is kernel function. The kernel function can transform low-dimensional nonlinear problems to high-dimensional programming linear problems. There are different kernels (linear kernel, polynomial kernel, radial basis function kernel, etc.) for performing tasks in high-dimensional feature spaces (Zhong et al. 2019).
Random forest
The RF is an extended variant of bagging (Bbeiman 1996). Bagging based on bootstrap sampling is the most famous parallel integrated learning methods (Efron and Tibshirani 1993). Bagging is a combination of bootstrapping and aggregating. Bootstrap sampling was performed on a data set of N samples to obtain dataset D1. The training model was repeated on D1 for M times to obtain M models, then the variance of the model could be reduced by averaging.
The base learner of RF is a decision tree. The selection of random attributes is further increased in the training process of the decision tree. RF is simple, easy to implement, has low computational overhead, and has powerful performance in many tasks. It is known as the method that represents integrated learning technology. The diversity of basic learners in RF not only comes from sample interference but also attribute interference. So the performance can be improved by increasing the divergence between each learner (Zhou 2016).
Extremely randomized trees
Also known as extra-trees (ET; Geurts et al. 2006), extremely randomized trees are tree-based randomization ensembles that combine the attribute randomization of a random subspace with a totally random selection of the cut-point. Compared to RF, it splits nodes by completely randomly selecting tangent points, and uses the entire learning sample to grow the tree. It can reduce the variance while slightly increasing the bias at the same time.
Gradient boosting machine
The GB (Friedman 2001) is a model with iterative stage addition that can be regarded as minimizing the steepest descent of a given loss function. GB is another ensemble method that can build a more powerful model by merging multiple decision trees. Unlike the RF method, GB uses a continuous method to construct trees, where each tree can reduce the error of the previous tree. There is no randomization in a GB regression tree, but strong pre-pruning is used. GB trees usually use trees with small depths so that the model occupies less memory and the prediction speed is faster.
Light gradient boosting machine
LightGBM (Ke et al. 2017) is a GB comprising reduce data dimensionality adaptation algorithm, as well as speed up process system. LightGBM is optimized by using: A decision tree algorithm based on a histogram, a leaf-wise leaf growth strategy with depth limitation, acceleration of histogram difference, and direct support for categorical features. It supports high-efficiency parallel training and has the advantages of fast training speed, low memory consumption, good accuracy and distributed support, and can rapidly process massive datasets.
Extreme gradient boosting
XGB (Chen and Guestrin 2016) is designed to be a highly scalable and accurate tree-boosting system. XGB incorporates a set of decision trees to build a powerful regression model. This large-scale machine learning method can easily and automatically apply multi-threaded parallelism to shorten execution times. In contrast to the GB model, XGB uses second-order Taylor expansion of the loss function. Additionally, the depth of the tree and the weight of the leaf nodes are part of the XGB objective function. It can reduce the iteration process and enhance the performance of the tree. A step-by-step decision tree growth technique is implemented to reduce model complexity.
Stacked generalization ensemble model
Stacked generalization (Wolpert 1992; Breiman 1996; Wolpert and Macready 1996) is the most popular meta-learning algorithm (Wolpert and Macready 1996; Sill et al. 2009; Zhang et al. 2010; Dou et al. 2020). Algorithm 1 shows the pseudo-code of the stacked generalization ensemble algorithm. Stacked generalization refers to any scheme for feeding information from one set of generalizers to another before forming the final guess. Stacked generalization can reduce the deviation of the generalizer from the provided learning set. The stacked generalization error is comprised of a term that depends on the generalization error of the individual learners and another term that contains all the correlations between learners, and is defined as:
where \(\overline{E }\) is the generalization errors of the individual learners, which depend on the errors of the individual learners Ei and the combined strategy algorithm; and \(\overline{A }\) is the ambiguities, which depend on the correlation between the individual learners Ai and the combined strategy algorithm (Krogh and Vedelsby 1995). From (3), it is obvious that increasing the ambiguity and decreasing individual generalization errors will improve the overall generalization. The errors of individual learners are small and the correlation between them is low, so the stacked model will be more accurate than individual learners. If a good algorithm is used to combine different individual learners to minimize the generalization error of the individual learners and maximize the ambiguities, then we can obtain the best stacked model.
The stacked algorithm structure of this article, shown in Fig. 3, uses the original dataset to train the primary learner. To avoid overfitting, cross-validation is used to train the primary learner. Then, the output of the primary learner is used as the new input features and the corresponding original tag is used as the new tag. In the first level, there are five learners: RF, ET, SVR, LightGBM and GB. First, fivefold cross-validation is used to train the training set, which generates new training data from each fold and then synthesizes a new training dataset from each single model. The new test dataset is the average of each fold model’s value, which is predicted by the original test dataset. In the second level, the new training dataset is trained by the XGB and then the final predicted value is obtained.
Particle swarm optimization
After the drilling rate model is built, we use algorithms to optimize the effective parameters and obtain the optimal ROP. In the oil and gas industry, many algorithms are used for difficult optimization problems, including gene expression programming, genetic programming, PSO, cuckoo optimization algorithm, biogeography-based optimizer, and imperialist competitive algorithm, social spider optimization, sine cosine algorithm, multi-verse optimization and moth flame optimization (Bodaghi et al. 2015; Hajirezaie et al. 2015, 2017a, b, 2019; Anemangely et al. 2018; Ashrafi et al. 2019; Sabah et al. 2019; Zhou et al. 2021). PSO resembles a school of flying birds and is an extremely simple and effective algorithm (Kennedy and Eberhart 1995). It requires only simple mathematical operators and is computationally inexpensive in terms of both memory requirements and speed. Hence, we use PSO to optimize ROP. The process for implementing PSO is as in Algorithm 2. In PSO, the particles are placed in the search space of the function and each particle evaluates the objective function at its current position. These particles move via cooperation and competition between the particles themselves. The position of the moved particle is expressed by the following equation:
where \({\overrightarrow{v}}_{i}\) is the rate of positional change, w is the inertial weight, c1 and c2 are two positive constants, rand1() and rand2()are two random functions in the range [0,1], \({\overrightarrow{p}}_{i}\) is the previous personal best position, \({\overrightarrow{p}}_{g}\) is the best position among all particles, and \({\overrightarrow{x}}_{i}\) is the current position (Shi and Eberhart 1998). We then evaluate the objective function at its current position and update \({\overrightarrow{p}}_{i}\) and\({\overrightarrow{p}}_{g}\). We keep updating the particle position and evaluation repeatedly until the end condition is met. Eventually, the swarm as a whole, like a flock of birds collectively foraging for food, is likely to move close to an optimum of the fitness function (Poli et al. 2007).
Results and discussion
Model performance
We use the leave-one-out method to verify model performance. The whole dataset is split, with 80% (1986 points) used as the training dataset and 20% (477 points) used as the testing dataset. We use R2, root mean square error (RMSE) and Mean Absolute Percentage Error (MAPE) to evaluate the performance of the model. The closer the R2 score is to 1, the better the goodness of fit of the model. It is expressed as:
where yi is the predicted values, xi is actual values. The RMSE is a measure of the spread of actual x values around the average of predicted y values. The smaller the RMSE value is, the higher the prediction accuracy. It is expressed as:
The machine learning algorithms were implemented using the scikit-learn library (Pedregosa et al. 2011). The tuning hyperparameters of these estimators were chosen using grid-search cross-validation of the training data. The performance and hyperparameters of each single model and stacked generalization model are shown in Table 3. The performance of the stacked generalization model is better than those of each single model. In the testing dataset, the R2 value of the stacked generalization model is 0.9568, higher than that of the best single model, ET. Meanwhile, the RMSE of the stacked generalization model is 0.4853 m/h, lower than that of the best model, ET. Figure 4 presents a regression plot of the predicted values versus the denoised values for each single model and stacked generalization model. Figure 5 presents the RMSE and R2 values of the different models in the testing dataset. Figure 6 presents the error of the predicted values versus the denoised values in the testing dataset. The error is expressed as:
Obviously, the regression coefficient values of the stacked generalization model are highly desirable, indicating that this model can make good predictions of drilling ROP.
The combination of learners brings three benefits (Dietterich 2000). First of all, from a statistical perspective, there can be more than one hypothesis that get the equal performance at the training set due to the fact that the hypothesis space is large. If we misselect a single learner at this time, we may get a poor generalization performance. Combining multiple learners will reduce this risk. Second, from a computational point of view, the algorithm often fall into a local minimum with poor generalization performance. This problem can be reduced by stacked generalization of multiple learner combinations. Third, from the perspective of representation, the true hypothesis of some learning tasks may not be within the hypothesis space considered by the current learning algorithm. At this time, using a single learner is invalid. By combining multiple learners, it is possible to learn a better approximation due to the expansion of the corresponding hypothesis space.
Learning algorithms that have problems due to statistical factors usually show high variance, while problems caused by calculation factors result in high calculation variance, and representation factors cause high deviations. Therefore, by combining several methods, the learning algorithm can reduce the impacts of variance and bias at the same time (Zhou 2016). Figure 7 shows the importance of the features in the ensemble models, RF, ET, GB and LightGBM. It can be seen that that the weight of each feature is different in each model. The top three import features are depth, porosity and FL in RF; depth, ECD and MW in ET; depth, porosity and ECD in GB; Q, WOB and MW in LightGBM. It indicates that the top three features which have greater influence on the prediction of target value of these models is different. The ambiguity between these learners is large and the accuracy of the individual model is high. Hence, the error can be reduced by combining these models. In this study, the XGB can combine different individual learners to minimize the generalization error of the individual learners and maximize the ambiguities, so the stacked generalization ensemble model can improve the accuracy.
Rate of penetration optimization
The stacked generalization ensemble model obtained in the previous section is incorporated into the optimization algorithm to optimize the parameters that maximize the ROP. Since depth increases during drilling, it was not considered as a decision variable. It makes sense that the parameters cannot be optimized for each meter of penetration. (Zhao et al. 2020). Therefore, we selected 23 points, one every 100 m, at which to optimize the ROP parameters to demonstrate the value of the proposed model. The selected parameters based on the formation and drilling design were T, bit diameter, and porosity. The critical parameters for downhole pressure and drilling safety are MW, ECD, and SPP, while FL is the key factor in protecting reservoirs and inhibiting water sensitivity. Therefore, MW, ECD, SPP, and FL cannot be changed arbitrarily. Hence, measured values of the parameters depth, T, bit diameter, porosity, MW, ECD, SPP, and FL are input into the model, while WOB, RPM, Q, and solid content are optimized via PSO to obtain the optimal ROP. A total of 20 particles are set; the first particle is set as the initial measured value of the drilling process, with other particles generated randomly. The search space of the optimized parameters is shown in Table 4. The optimization results are shown in Fig. 8; the ROP is better than the initial value, with the mean ROP increasing by 33.5% from 3.25 to 4.34 m/h.
Conclusions
This work presents a stacked generalization ensemble model for predicting the ROP while drilling. The model combines six efficient methods: SVR, RF, ET, GB, LightGBM and XGB. The stacked generalization ensemble model has two levels. In the first level, there are five learners: RF, ET, SVR, LightGBM and GB. New training data are generated through fivefold cross-validation and the original target value is used as the target value. In the second level, the new training dataset is trained by XGB. The ambiguity between these learners of first level is large and the accuracy of the individual model is high. The XGB can combine different individual learners to minimize the generalization error of the individual learners and maximize the ambiguities. The performance of the stacked model is better than each single model, the R2 value is 0.9568 and the RMSE is 0.4853 m/h. The model provides a method for predicting drilling rates with high accuracy, which is beneficial to the optimization of ROP. By optimizing the effective drilling parameters via PSO at the selected 23 points, the mean ROP can be increased by 33.5%.
References
Abbas AK, Rushdi S, Alsaba M, Al Dushaishi MF (2019b) Drilling rate of penetration prediction of high-angled wells using artificial neural networks. J Energy Resour Technol Trans ASME 141:1–11. https://doi.org/10.1115/1.4043699
Abbas AK, Rushdi S, Alsaba M (2019a) Modeling rate of penetration for deviated wells using artificial neural network. In: Society of petroleum engineers - Abu Dhabi international petroleum exhibition and conference 2018, ADIPEC 2018
Ahmed AI, Ibrahim AA (2019) Bourgoyne and young model development review. Int J Eng Sci Res Technol 8:164–174
Ahmed OS, Adeniran AA, Samsuri A (2019) Computational intelligence based prediction of drilling rate of penetration: a comparative study. J Pet Sci Eng 172:1–12. https://doi.org/10.1016/j.petrol.2018.09.027
Al-Abduljabbar A, Elkatatny S, Mahmoud M et al (2019) A robust rate of penetration model for carbonate formation. J Energy Resour Technol Trans ASME 141:1–9. https://doi.org/10.1115/1.4041840
Anemangely M, Ramezanzadeh A, Tokhmechi B et al (2018) Drilling rate prediction from petrophysical logs and mud logging data using an optimized multilayer perceptron neural network. J Geophys Eng 15:1146–1159. https://doi.org/10.1088/1742-2140/aaac5d
Ansari HR, Sarbaz Hosseini MJ, Amirpour M (2017) Drilling rate of penetration prediction through committee support vector regression based on imperialist competitive algorithm. Carbonates Evaporites 32:205–213. https://doi.org/10.1007/s13146-016-0291-8
Arabjamaloei R, Karimi Dehkordi B (2012) Investigation of the most efficient approach of the prediction of the rate of penetration. Energy Sources Part A Recover Util Environ Eff 34:581–590. https://doi.org/10.1080/15567036.2010.493925
Arabjamaloei R, Shadizadeh S (2011) Modeling and optimizing rate of penetration using intelligent systems in an Iranian southern oil field (ahwaz oil field). Pet Sci Technol 29:1637–1648. https://doi.org/10.1080/10916460902882818
Arabjamaloei R, Edalatkhah S, Jamshidi E (2011) A new approach to well trajectory optimization based on rate of penetration and wellbore stability. Pet Sci Technol 29:588–600. https://doi.org/10.1080/10916460903419172
Ashrafi SB, Anemangely M, Sabah M, Ameri MJ (2019) Application of hybrid artificial neural networks for predicting rate of penetration (ROP): a case study from Marun oil field. J Pet Sci Eng 175:604–623. https://doi.org/10.1016/j.petrol.2018.12.013
Bahari A, Baradaran Seyed A (2007) Trust-region approach to find constants of Bourgoyne and Young penetration rate model in Khangiran Iranian gas field. In: Latin American & Caribbean petroleum engineering conference. Society of Petroleum Engineers
Bahari MH, Bahari A, Nejati F, et al (2009) Drilling rate prediction using bourgoyne and young model associated with genetic algorithm
Basarir H, Tutluoglu L, Karpuz C (2014) Penetration rate prediction for diamond bit drilling by adaptive neuro-fuzzy inference system and multiple regressions. Eng Geol 173:1–9. https://doi.org/10.1016/j.enggeo.2014.02.006
Bbeiman LEO (1996) Bagging predictors. Mach Learn 140:123–140
Bello O, Teodoriu C, Yaqoob T, et al (2016) Application of artificial intelligence techniques in drilling system design and operations: a state of the art review and future research pathways. In: Soc Pet Eng - SPE Niger Annu Int Conf Exhib. https://doi.org/10.2118/184320-ms
Bezminabadi SN, Ramezanzadeh A, Jalali S-ME et al (2017) Effect of rock properties on ROP modeling using statistical and intelligent methods: a case study of an oil well in southwest of Iran. Arch Min Sci 62:131–144
Bingham M. (1965) A new approach to interpreting rock drillability. Re-printed from Oil Gas J
Bodaghi A, Ansari HR, Gholami M (2015) Optimized support vector regression for drilling rate of penetration estimation. Open Geosci 7:870–879. https://doi.org/10.1515/geo-2015-0054
Bourgoyne AT, Young FS (1974) A multiple regression approach to optimal drilling and abnormal pressure detection. Soc Pet Eng J 14:371–384. https://doi.org/10.2118/4238-PA
Breiman L (1996) Stacked regressions. Mach Learn 24:49–64. https://doi.org/10.1023/A:1018046112532
Breiman L (2001) Random forests. Mach Learn 45:5–32
Chen T, Guestrin C (2016) XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery, New York, NY, USA, pp 785–794
Diaz MB, Kim KY, Shin HS, Zhuang L (2019) Predicting rate of penetration during drilling of deep geothermal well in Korea using artificial neural networks and real-time data collection. J Nat Gas Sci Eng 67:225–232. https://doi.org/10.1016/j.jngse.2019.05.004
Dietterich TG (2000) Ensemble methods in machine learning. In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Springer Verlag, pp 1–15
Dou J, Yunus AP, Bui DT et al (2020) Improved landslide assessment using support vector machine with bagging, boosting, and stacking ensem- ble machine learning framework in a mountainous watershed. Japan. https://doi.org/10.1007/s10346-019-01286-5
Efron B, Tibshirani RJ (1993) An Introduction to the bootstrap. Chapman & Hall, New York
Elkatatny S (2018) New approach to optimize the rate of penetration using artificial neural network. Arab J Sci Eng 43:6297–6304. https://doi.org/10.1007/s13369-017-3022-0
Elkatatny S (2019) Development of a new rate of penetration model using self-adaptive differential evolution-artificial neural network. Arab J Geosci. https://doi.org/10.1007/s12517-018-4185-z
Elkatatny S, Al-abduljabbar A, Abdelgawad K (2020) A new model for predicting rate of penetration using an artificial neural network. Sensors. https://doi.org/10.3390/s20072058
Eskandarian S, Bahrami P, Kazemi P (2017) A comprehensive data mining approach to estimate the rate of penetration: application of neural network, rule based models and feature ranking. J Pet Sci Eng 156:605–615. https://doi.org/10.1016/j.petrol.2017.06.039
Fletcher R (2013) Practical methods of optimization. Wiley
Friedman JH (2001) Greedy function approximation: a gradient boosting machine. Ann Stat 29:1189–1232. https://doi.org/10.1214/aos/1013203451
Gan C, Cao W, Wu M et al (2019a) Two-level intelligent modeling method for the rate of penetration in complex geological drilling process. Appl Soft Comput J 80:592–602. https://doi.org/10.1016/j.asoc.2019.04.020
Gan C, Cao WH, Wu M et al (2019b) Prediction of drilling rate of penetration (ROP) using hybrid support vector regression: a case study on the Shennongjia area. Central China J Pet Sci Eng. https://doi.org/10.1016/j.petrol.2019.106200
Garcia LPF, De CACPLF, Lorena AC (2015) Neurocomputing Effect of label noise in the complexity of classi fi cation problems. Neurocomputing 160:108–119. https://doi.org/10.1016/j.neucom.2014.10.085
Geurts P, Ernst D, Wehenkel L (2006) Extremely randomized trees. Mach Learn 63:3–42. https://doi.org/10.1007/s10994-006-6226-1
Hajirezaie S, Hemmati-Sarapardeh A, Mohammadi AH et al (2015) A smooth model for the estimation of gas/vapor viscosity of hydrocarbon fluids. J Nat Gas Sci Eng 26:1452–1459. https://doi.org/10.1016/J.JNGSE.2015.07.045
Hajirezaie S, Pajouhandeh A, Hemmati-Sarapardeh A et al (2017a) Development of a robust model for prediction of under-saturated reservoir oil viscosity. J Mol Liq 229:89–97. https://doi.org/10.1016/J.MOLLIQ.2016.11.088
Hajirezaie S, Wu X, Peters CA (2017b) Scale formation in porous media and its impact on reservoir performance during water flooding. J Nat Gas Sci Eng 39:188–202. https://doi.org/10.1016/J.JNGSE.2017.01.019
Hajirezaie S, Wu X, Soltanian MR, Sakha S (2019) Numerical simulation of mineral precipitation in hydrocarbon reservoirs and wellbores. Fuel 238:462–472. https://doi.org/10.1016/J.FUEL.2018.10.101
Hareland G, Rampersad PR (1994) Drag - bit model including wear. SPE Lat Am Pet Eng Conf 11
Hegde C, Gray KE (2017) Use of machine learning and data analytics to increase drilling efficiency for nearby wells. J Nat Gas Sci Eng 40:327–335. https://doi.org/10.1016/j.jngse.2017.02.019
Hegde C, Gray K (2018) Evaluation of coupled machine learning models for drilling optimization. J Nat Gas Sci Eng 56:397–407. https://doi.org/10.1016/j.jngse.2018.06.006
Hegde C, Daigle H, Millwater H, Gray K (2017) Analysis of rate of penetration (ROP) prediction in drilling using physics-based and data-driven models. J Pet Sci Eng 159:295–306. https://doi.org/10.1016/j.petrol.2017.09.020
Hegde C, Wallace S, Gray K (2015) Using trees, bagging, and random forests to predict rate of penetration during drilling. In: Soc Pet Eng - SPE Middle East Intell Oil Gas Conf Exhib. https://doi.org/10.2118/176792-ms
Hegde C (2016) Application of statistical learning techniques for rate of penetration (ROP) prediction in drilling
Hua Z (2010) Application of Bourgoyne and Young penetration rate prediction model in gas drilling. J Chongqing Univ Ence Technol Ences Ed 13:6–7
Kahraman S (2016) Estimating the penetration rate in diamond drilling in laboratory works using the regression and artificial neural network analysis. Neural Process Lett 43:523–535. https://doi.org/10.1007/s11063-015-9424-7
Ke G, Meng Q, Finley T, et al (2017) Lightgbm: a highly efficient gradient boosting decision tree. In: Advances in neural information processing systems. pp 3146–3154
Kennedy J, Eberhart R (1995) Particle swarm optimization. In: Proceedings of ICNN’95 - International Conference on Neural Networks. pp 1942–1948 vol.4
Krogh A, Vedelsby J (1995) Neural network ensembles, cross validation, and active learning. Adv Neural Inf Process Syst 7:231–238
Mantha B, Samuel R (2016) ROP optimization using artificial intelligence techniques with statistical regression coupling. In: Proceedings - SPE annual technical conference and exhibition
Maurer MC (1962) The “Perfect - Cleaning” theory of rotary drilling. J Pet Technol 14:1270–1274
Motahhari HR, Hareland G, James JA (2010) Improved drilling efficiency technique using integrated PDM and PDC bit parameters. J Can Pet Technol 49:45–52. https://doi.org/10.2118/141651-PA
Motahhari HR (2008) Improved drilling efficiency technique using integrated PDM and PDC Bit Parameters. University of Calgary
Nascimento A, Tamas Kutas D, Elmgerbi A et al (2015) Mathematical modeling applied to drilling engineering: an application of Bourgoyne and Young ROP model to a presalt case study. Math Probl Eng. https://doi.org/10.1155/2015/631290
Orr K (1998) Data quality and systems theory. Commun ACM 41:66–71
Pedregosa F, Varoquaux G, Gramfort A et al (2011) Scikit-learn: machine learning in python. J Mach Learn Res 12:2825–2830
Poli R, Kennedy J, Blackwell T (2007) Particle swarm optimization. Swarm Intell 1:33–57
Qian L, Cao Y, Zhu H (2021) Discussion on the lower limit of data validity for ROP prediction based on artificial intelligence. Drill Eng 48:21–30. https://doi.org/10.12143/j.ztgc.2021.03.003
Rahimzadeh H, Mostofi M, Hashemi A (2011) A new method for determining Bourgoyne and Young penetration rate model constants. Pet Sci Technol 29:886–897
Redman TC (1998) The impact of poor data quality on the typical enterprise. Commun ACM 41:79–82
Sabah M, Talebkeikhah M, Wood DA et al (2019) A machine learning approach to predict drilling rate using petrophysical and mud logging data. Earth Sci Inf 12:319–339. https://doi.org/10.1007/s12145-019-00381-4
Savitzky A, Golay MJE (1964) Smoothing and differentiation of data by simplified least squares procedures. Anal Chem 36:1627–1639. https://doi.org/10.1021/ac60214a047
Shi X, Liu G, Gong X et al (2016) An efficient approach for real-time prediction of rate of penetration in offshore drilling. Math Probl Eng. https://doi.org/10.1155/2016/3575380
Shi Y, Eberhart R (1998) A modified particle swarm optimizer. In: 1998 IEEE International Conference on Evolutionary Computation Proceedings. IEEE world congress on computational intelligence (Cat. No.98TH8360). pp 69–73
Sill J, Takács G, Mackey L, Lin D (2009) Feature-weighted linear stacking. arXiv Prepr arXiv09110460
Soares C, Gray K (2019) Real-time predictive capabilities of analytical and machine learning rate of penetration (ROP) models. J Pet Sci Eng. https://doi.org/10.1016/j.petrol.2018.08.083
Soares C, Daigle H, Gray K (2016) Evaluation of PDC bit ROP models and the effect of rock strength on model coefficients. J Nat Gas Sci Eng 34:1225–1236. https://doi.org/10.1016/j.jngse.2016.08.012
Sun J, Li Q, Chen M et al (2019) Optimization of models for a rapid identification of lithology while drilling - A win-win strategy based on machine learning. J Pet Sci Eng 176:321–341. https://doi.org/10.1016/j.petrol.2019.01.006
Vapnik VN (1995) The nature of statistical learning theory. Springer, New York, NY
Vapnik VN, Chervonenkis A (1964) A note on one class of perceptrons. Autom Remote Control 25:821–837
Warren TM (1987) Penetration-rate performance of roller-cone bits. SPE Drill Eng 2:9–18. https://doi.org/10.2118/13259-PA
Wolpert DH (1992) Stacked generalization. Neural Netw 5:241–259. https://doi.org/10.1016/S0893-6080(05)80023-1
Wolpert D, Macready W (1996) Combining stacking with bagging to improve a learning algorithm
Zhang X, Zhou D, Wang L (2010) Stacking algorithms for automated container ports: An improvement by direct stacking. In: Proceedings - 2010 2nd WRI Global Congress on Intelligent Systems, GCIS 2010. IEEE, pp 35–38
Zhao Y, Noorbakhsh A, Koopialipoor M et al (2020) A new methodology for optimization and prediction of rate of penetration during drilling operations. Eng Comput. https://doi.org/10.1007/s00366-019-00715-2
Zhong R, Johnson RL, Chen Z (2019) Using machine learning methods to identify coals from drilling and logging-while-drilling LWD data. SPE/AAPG/SEG Asia Pacific Unconv Resour Technol Conf 2019, APUR 2019
Zhou Z (2016) Machine learning. Tsinghua University Press, Beijing
Zhou J, Qiu Y, Armaghani DJ et al (2021) Predicting TBM penetration rate in hard rock condition: a comparative study among six XGB-based metaheuristic techniques. Geosci Front 12:101091. https://doi.org/10.1016/j.gsf.2020.09.020
Funding
This work was supported by the National Natural Science Foundation of China under Grant 61733016, No.1 Institute Geology and Mineral Resources of Shandong Province under Grant 2020DW01, the Qinghai Province Key R&D and Transformation Program under Grant 2020-SF-149, the National Key R&D Program of China under Grant 2018YFC0603405.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
On behalf of all the co-authors, the corresponding author states that there is no conflict of interest.
Data availability
https://github.com/liunaipeng/rop/blob/main/qcrop.csv
Code availability
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Liu, N., Gao, H., Zhao, Z. et al. A stacked generalization ensemble model for optimization and prediction of the gas well rate of penetration: a case study in Xinjiang. J Petrol Explor Prod Technol 12, 1595–1608 (2022). https://doi.org/10.1007/s13202-021-01402-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13202-021-01402-z