Literature DB >> 33458545

Real-Time Prediction of Equivalent Circulation Density for Horizontal Wells Using Intelligent Machines.

Ahmed Alsaihati1, Salaheldin Elkatatny1, Abdulazeez Abdulraheem1.   

Abstract

Equivalent circulation density (ECD) is an important part of drilling fluid calculations. Analytical equations based on the conservation of mass and momentum are used to determine the ECD at various depths in the wellbore. However, these equations do not incorporate important factors that have a direct impact on the ECD, such as bottom-hole temperature, pipe rotation and eccentricity, and wellbore roughness. This work introduced different intelligent machines that could provide a real-time accurate estimation of the ECD for horizontal wells, namely, the support vector machine (SVM), random forests (RF), and a functional network (FN). Also, this study sheds light on how principal component analysis (PCA) can be used to reduce the dimensionality of a data set without loss of any important information. Actual field data of Well-1, including drilling surface parameters and ECD measurements, were collected from a 5-7/8 in. horizontal section to develop the models. The performance of the models was assessed in terms of root-mean-square error (RMSE) and coefficient of determination (R 2). Then, the best model was validated using unseen data points of 1152 collected from Well-2. The results showed that the RF model outperformed the FN and SVM in predicting the ECD with an RMSE of 0.23 and R 2 of 0.99 in the training set and with an RMSE of 0.42 and R 2 of 0.99 in the testing set. Furthermore, the RF predicted the ECD in Well-2 with an RMSE of 0.35 and R 2 of 0.95. The developed models will help the drilling crew to have a comprehensive view of the ECD while drilling high-pressure high-temperature wells and detect downhole operational issues such as poor hole cleaning, kicks, and formation losses in a timely manner. Furthermore, it will promote safer operation and improve the crew response time limit to prevent undesired events.
© 2020 The Authors. Published by American Chemical Society.

Entities:  

Year:  2020        PMID: 33458545      PMCID: PMC7808159          DOI: 10.1021/acsomega.0c05570

Source DB:  PubMed          Journal:  ACS Omega        ISSN: 2470-1343


Introduction

Drilling fluid hydraulics has played an important role in well design when drilling a vertical or extended reach well. Therefore, accurate model and optimized drilling fluid hydraulics are crucial to allow engineers to properly design a well profile, thereby improving drilling efficiency, reducing risks, and decreasing nonproductive time (NPT). Additionally, it empowers the drilling crew to fully examine the downhole conditions and identify potential problems such as drill string washout, plugged nozzles, and the presence of a gas kick in the well. Equivalent circulation density (ECD) is an important aspect of drilling fluid hydraulics which helps in avoiding kicks and drilling fluid losses, particularly in deep high-pressure high-temperature (HPHT) wells, where the temperature is significant and the margin between pore pressure and fracture pressure is minute.[1] Drilled cuttings can increase the effective drilling fluid density and reduce fluid flow area, which in turn increases the value of the ECD. Thus, ECD calculations can be used as a baseline to monitor hole cleaning in real time while drilling. Hydraulics programs have been widely used to evaluate the detailed hydraulic calculations that are required in planning, execution, and post-well review phases of drilling a well. The user has the option to select which rheological model (Bingham plastic, Power-law, Herschel–Bulkley) the hydraulic calculations are based on. Each rheological model, however, requires different data inputs, which can be obtained from a full lab test. These parameters are then fed into the software to compute the drilling fluid hydraulics and its associated parameters such as ECD. It has been observed that there is a discrepancy between computed data and the data recorded in the field.[2] One factor contributing to the inconsistency between calculated and actual drilling hydraulics is the assumption that the rheological properties of the drilling fluid are independent of pressure and temperature.[1] This can be valid in shallow wells, where temperature changes are insignificant, and hence, the rheological variations are small. Furthermore, the mathematical equations used to calculate the drilling hydraulics entail a set of assumptions such as[2−4] (i) concentric annular and circular sections, (ii) laminar and turbulent flow, where plug flow is considered as laminar flow, while transition flow is neglected, and (iii) steady-state flow, where fluid properties at any single point in the system do not change. Such potentially unrealistic assumptions cannot be fulfilled in all drilling conditions.[5] In HPHT wells, evaluations and analyses of the effect of temperature and pressure on drilling hydraulics and kick probability are necessary.[1,6] Rommeveit and Bjorkevoll[1] developed two models, a static ECD model and a dynamic ECD model, which incorporate the temperature profile along the wellbore and allow the mud properties to be dependent on pressure and temperature. They concluded that the models made it possible to make realistic and reliable evaluations of any operational concerns during drilling especially when, for example, pit gain and standpipe pressure (SPP) deviate from the predictable value. Scheid et al.[7] performed an experiment to evaluate the frictional pressure loss through pipes, annuli, and accessories such as pipe tool joint and stabilizers to estimate the drilling hydraulic calculations and the corresponding ECD with four types of drilling fluid. They stated that the results obtained in the study can be used in the drilling industry for accurate drilling hydraulic calculations. Dokhani et al.[3] concluded that eccentricity (ϵ), which describes how off-center a pipe is in an open hole or casing, affects frictional pressure loss and hence the ECD for Herschel–Bulkley fluid. Eccentricity reduces overall frictional pressure loss if ϵ is >0.1 and it will be neglected if ϵ is 0.8.[8−10] Moreover, Dokhani et al.[3] evaluated the effect of pipe rotations on frictional pressure loss. They observed that a drilling fluid’s apparent viscosity decreases as pipe rotation increases; hence, overall frictional loss decreases. This phenomenon is a result of the shear-thinning properties of Herschel–Bulkley fluid.[11] Dokhani et al.[3] also recommended considering the effect of roughness of the wellbore wall to have a more accurate estimation of drilling hydraulic calculations such as ECD. A more accurate and reliable way to evaluate the ECD is to use a downhole pressure sensor, which consists of high-accuracy pressure gauges to measure the annular pressure. The sensors provide important real-time downhole pressure information which will allow the drilling crew to make a faster and better decision.[12] However, such sensors are extremely expensive.[13] The literature review shows that major discrepancies exist between actual drilling hydraulic values and those predicted by previously accepted mathematical equations. In addition, several studies have suggested consideration of other factors, including pipe eccentricity, wellbore roughness, pressure and temperature, and pipe rotation speed to improve the accuracy of drilling hydraulic calculations. This, however, would mandate intensive mathematical intervention. An alternative approach to estimating the ECD with higher accuracy is by using artificial intelligence (AI) and machine learning (ML),[13,14] which map the inputs and outputs based on a defined algorithm.[15] AI is efficient enough to minimize human efforts in various areas and does not necessitate fundamental knowledge of physics and science. AI has become an important subject in the drilling industry,[15] where streaming data are continuously structured from the surface, downhole sensors and logging. AI and ML allow for new methods to learn from these data to mitigate drilling challenges, describe trends in real time, and automate drilling.[16] This study introduces three intelligent techniques to predict the ECD which can help in reducing the overall cost of drilling operation by not using downhole pressure sensors for future offset wells. The developed models would give the drilling crew the capability to monitor hole cleaning, maintain the wellbore pressures in horizontal extended-reach wells in real-time, and reduce the risk of formation fracture and collapse. Furthermore, they can improve the warning time to detect kicks to maintain well control.

Application of AI and ML in the Drilling Industry

Drilling is a major aspect of the oil and gas industry which is considered to be highly expensive and risky. Therefore, tools that can improve the drilling operation at a low cost are essential. There are some applications related to AI and ML in drilling operations such as[17] well planning, rate of penetration (ROP) optimization, well integrity, detecting problems, procedure decision making, and pattern recognition. The large number of publications on the application of AI and ML in the drilling industry indicate that AI and ML can potentially reduce drilling costs and promote safety at the rig-site.[16,17] Abdelgawad et al.[13] used SPP, ROP, and mud weight (MW) as input parameters to predict the ECD using an artificial neural network (ANN) and an adaptive neuro-fuzzy inference system (ANFIS) in an 8–1/2 in. vertical hole section. The two models predicted the ECD with a correlation coefficient (R) of 0.99 and AAPE of 0.22%. Alkinani et al.[14] collected data from more than 2000 wells located around the world and used an ANN to build a model to predict the ECD. The input parameters for the model were flow rate, MW, nozzles total flow area, plastic velocity, revolutions per minute, weight on bit (WOB), and yield point. Bayesian Regularization (BR) was used as a training algorithm because it had the highest R2 (0.982). Alsabaa et al.[18] used ANFIS to predict the rheological properties of invert emulsion mud in real time. Al-Yami et al.[19] used Bayesian Belief Network (BBN) to establish an intelligent drilling system based on different fluid and reservoir properties. This tool can be utilized to train young engineers in different drilling perspectives such as well control, underbalanced drilling, drilling fluid, and cementing best practices. Ahmadi[20] simulated the performance of various types of drilling fluid’s rheology under different conditions using a support vector machine (SVM) with good agreement between lab results and prediction. Elkatatny et al.[21] used an ANN, which incorporates drilling fluid and drilling mechanical properties, to predict ROP with high accuracy. Elkatatny et al.[22] used actual field data to build an ANN to predict the top-depth of four geological formations in real time while drilling. The best results were obtained when scaled conjugate gradient backpropagation (TRAINSCG) with 20 neurons was used. The input parameters used in the study to develop the model were the mechanical drilling parameters including ROP, Q, RS, SPP, torque, and WOB.

Support Vector Machines

SVMs are supervised ML models that analyze data for classification or regression problems[17] and can lead to a high performance in particular applications.[23] Moreover, SVMs have advantages over other ML algorithms, including generalization capability, less learning time, and strong interference capacity.[24,25] SVMs move the data from a low dimension to a high dimension, denoted as kernel space, to find a support vector classifier, that is, hyperplane, that minimizes the number of misclassified data points.[17,26,27] The hyperplane is defined as the plane with a maximal margin of separation between two classes, and the distance between the hyperplane and the nearest data point from either set is known as the margin. Support vectors are the data with the closest distance to the hyperplane. These are difficult to classify,[26] which is why the hyperplane with the maximum margin is the best separator. To transform the data to kernel space, SVMs use kernel functions to systematically find the support vector classifiers in higher dimensions. When kernels are used to transform the feature vectors from input space to kernel space for linearly nonseparable data sets, the kernel matrix computation requires computational resources. The popular kernel functions are[27] linear kernel function, polynomial kernel function, Gaussian RBF, and randomized blocks analysis of variance (ANOVA RB) kernel. The selection of kernel functions is essentially dependent on the nature of the data set. The linear kernel ranks behind the polynomial kernel, and it is useful in large sparse data vectors. On the other hand, the polynomial kernel is commonly used in image processing. While the ANOVA RB kernel is generally used for regression tasks, the Gaussian RBF is mostly applied if the user lacks prior knowledge.[27]

Random Forests

The RF is an ensemble learning technique that can be used for classification and regression problems.[28] The RF combines hundreds or thousands of decision trees and trains each one on a slightly different set of observations. It splits nodes in each tree considering a limited number of features, in a process called bootstrapping.[29] The final prediction of the RF is made by averaging the predictions of each tree in a process called aggregation. When a bootstrap sample is made from the data set, about one-third of the samples from the original data set are not included.[30] These samples are called out-of-bag data and are used to measure internally the accuracy of the RF. The RF outperforms a single decision tree because of its ability to limit overfitting without substantially increasing the margin of error.[29]

Functional Network

A functional network (FN) is a generalization of an ANN, which can be accomplished by using multiple arguments and learnable functions, that is, in a FN, the activation functions associated with neurons are not fixed but learned from data.[31] In an ANN, the weights associated with the neurons must be learned, while they are suppressed in a FN.[25] Another characteristic of a FN is that the specification of the initial topology could be based on the features of the problem. Therefore, knowledge about the problem can help in developing the network structure.

Principal Component Analysis

PCA is a useful statistical technique that has many applications in many fields and is a common technique for finding patterns in a data set with high dimensions.[32] The objective of PCA is to reduce the dimensionality of a large data set of variables or features into a smaller one without the loss of any important information.[33] PCA finds a lower dimensional space (W) to transform the data (X = [x1, x2,..., x]) from a higher dimensional space (RM) to a lower dimensional space (Rk),where N represents the total number of observations (rows in a data set) and x represents the ith observation. Each observation is represented by M features or variables (columns in a dataset), that is, each observation is represented as a point in M-dimensional space.[34] A PCA space for a data set that contains a number of features K has K principal components. These K principal components are uncorrelated, orthonormal, and represent the direction of the maximum variance.[33] The first component, (PC1 ϵ RM), always represents the maximum variation of the data; (PC2 ϵ RM) represents the second-largest variation of the data, while (PC ϵ RM) represents the least variation of the data.[32,33] The principal components that represent more than 90% of the variation in a data set are often considered. The following sequence shows how to calculate PCs using the Covariance Matrix Method for a given data set (X = {x1, x2, ..., x}): Compute the mean of all variables Subtract the mean of each variable from all observations corresponding to that variable Compute the covariance matrix for the centered data set Compute the eigenvectors and eigenvalues of the covariance matrix Sort eigenvectors according to their corresponding eigenvalues Select the eigenvectors that have the largest eigenvalues. The selected eigenvectors represent the projection space of PCA Project all observations on the lower dimensional space of PCA

Methodology

Data Collections

Two types of actual field data of Well-1 were collected from a 5–7/8 in. horizontal section: (i) drilling surface parameters and (ii) ECDs. The drilling surface parameters including flow rate (Q), hook-load (HL), ROP, rotary speed (RS), SPP, WOB, and surface drilling torque (T) were obtained from surface real-time transmitter sensors, while ECDs were obtained from a pressure-while-drilling (PWD) sensor. A total of 3567 data points of ECD were obtained at the same depth of the drilling surface parameters. Table shows the statistical parameters of the whole data set. Q ranges from 250 to 296.5 GPM; HL ranges from 256 to 286 klbf; ROP ranges from 3.5 to 59.6 ft/h; RS ranges from 59 to 141 RPM; SPP ranges from 2354.8 to 3656.5 psi; WOB ranges from 5.1 to 20.1 kIbf; T ranges from 2.9 to 10 kft.Ibf; and the ECD ranges from 83.4 to 95.5 pcf.
Table 1

Statistical Parameters of the Whole Data set (3567 Data Points)

statistical parametersQ (gal/min)HL (klbfROP (ft/h)RS (RPM)SPP (psi)WOB (klbf)T (kft.lbf)ECD (Pcf)
minimum250.0256.03.559.02354.85.12.983.4
maximum296.5286.059.6141.03656.520.110.095.5
mean276.7267.423.0119.73031.715.26.990.4
standard deviation10.35.56.217.1257.73.01.23.2
skewness–1.660.620.18–0.95–0.15–0.96–0.04–0.37
kurtosis1.090.161.761.34–0.110.10–0.85–0.90
Figure compares the linear relationship of the input parameters used to train the model with the ECD. Figure shows that HL, SPP, and T have a strong relationship with the ECD, that is, 0.71, 0.87 and 0.85, respectively. Q has a moderate linear relationship with an ECD of −0.38. On the other hand, ROP, RPM, and WOB have relatively a low relationship of −0.13, 0.1, and 0.11, respectively.
Figure 1

Relative importance of the input variable set.

Relative importance of the input variable set.

Data Splitting

In ML, it is necessary to build a model that makes accurate predictions for future data. Thus, the data set is divided into two portions, training and testing sets. The training set is used to ensure that the machine recognizes patterns in the data set, while the testing set is used to evaluate how well the machine can predict unseen data based on its training. In this analysis, seven surface drilling parameters were used as independent variables (inputs): Q, HL, ROP, RS, SPP, WOB, and T, while ECD was used as a dependent variable (output). The data set was randomly split, with a ratio of 80:20. Eighty percent of the data was selected for training to ensure that the models capture most of the ECD variation while drilling with various surface drilling parameters. The training set has 2742 data points, while the testing set has 827 data points. Tables and 3 show the statistical parameters of the training and testing sets, respectively.
Table 2

Statistical Parameters of the Training Set (2742 Data Points)

statistical parametersQ (gal/minHL (klbf)ROP (ft/h)RS (RPM)SPP (psi)WOB (klbf)T (kft.lbf)ECD (Pcf)
minimum249.4256.13.559.02379.75.53.783.4
maximum296.6285.259.6141.33632.120.010.095.5
mean276.7267.423.0119.83035.315.26.990.4
standard deviation10.35.56.216.9258.03.01.23.2
skewness–1.670.610.22–0.93–0.15–0.96–0.05–0.39
kurtosis1.110.121.881.28–0.140.08–0.87–0.89
Table 3

Statistical Parameters of the Testing Set (827 Data Points)

statistical parametersQ (gal/min)HL (klbf)ROP (ft/h)RS (RPM)SPP (psi)WOB (klbf)T (kft.lbf)ECD (Pcf)
minimum250.3256.14.959.02354.85.13.283.4
maximum296.0286.156.7140.23656.520.110.095.5
mean276.6267.322.8119.43019.615.16.890.2
standard deviation10.35.65.917.7256.53.01.23.2
skewness–1.630.660.04–1.01–0.16–0.97–0.01–0.30
kurtosis1.040.311.241.460.010.15–0.80–0.94

Model Building

Python library’s Scikit-Learn was used to build the SVM model. The SVM parameters, known as hyperparameters (e.g., regularization parameter C, gamma, and kernel type), were tuned using a built-in function in Scikit-Learn known as GridSearchCV to evaluate the improvement and performance of the SVM. Two types of kernel, RBF and linear were tried while varying the value of C from 0.001 to 1000 and the type of gamma (i.e. auto and scale). Likewise, Scikit-Learn was used to develop the RF model. The model parameters, including the maximum depth of the tree “max_depth”, the maximum features to be considered when splitting the node in each tree “max_features”, and the number of the trees in the forest” n_estimators”, were tuned using GridSearchCV. Different values of max_depth from 3 to 21 and three types of max_features (i.e., auto, sqrt, and log2) were tried while varying the n_estimator from 3 to 150. The rest of the model parameters, including minimum_samples split, minimum_sample leaf, maximum_leaf nodes, and minimum_impurity, were kept as default values. MATLAB code was used to build the FN model. Two methods, FN Forward-Backward Method (FNFBM) and FN Exhaustive-Backward Method (FNEBM), were studied with two types of relationship: linear and nonlinear.

Using PCA for Dimensionality Reduction

Standardizing a data set refers to shifting the distribution of each variable to have a unit scale, that is, a mean of zero and a standard deviation of one, which is a necessary step for the PCA algorithm. The values of the input parameters were standardized using the following equationwhere Y is the normalized input parameter, X is the input parameter to be normalized, and σ is the standard deviation of the input variable. Then, Python library’s Scikit-Learn was used to apply the PCA algorithm to the input variables of the data set of Well-1. The transformation to PCA space was completed in three steps: (i) instantiate the PCA by passing the number of principal components to the constructor, (ii) call the fit, which will find the covariance matrix, the eigenvectors, and eigenvalues of the covariance matrix, and (iii) transform the data set into the PCA space. Then, the transformed data set was fed into another RF model for training and testing. The PCA-based RF model improvement and performance while using different numbers of PCs were evaluated.

Results and Discussion

Model Assessment

The SVM model performance with different parameters (C, gamma, and kernel type) is presented in Table . The optimal parameters for each kernel type were obtained using GridSearchCV. Table shows that the SVM model with the RBF kernel had the highest R2 and the lowest root-mean-square error (RMSE) compared to the linear kernel. The SVM with the RBF kernel predicted the ECD with an R2 of 0.97 and an RMSE of 0.54 in the training set, while the R2 and RMSE were 0.97 and 0.58, respectively, in the testing set. On the other hand, the SVM with the linear kernel predicted the ECD with an R2 of 0.95 and an RMSE of 0.71 in the training set and with an R2 and RMSE of 0.95 and 0.74, respectively, in the testing set. Figure a,b are cross-plots of the actual and predicted ECD of the training and testing sets for the SVM with the kernel type.
Table 4

Performance of the SVM Model with Different Kernel Types

kernel typeCgammaRMSE_trainingR2_trainingRMSE_testingR2_testing
linear0.001auto0.710.950.740.95
RBF500scale0.540.970.580.97
Figure 2

Cross-plot of the actual and predicted ECD with RBF kernel (a) training set and (b) testing set.

Cross-plot of the actual and predicted ECD with RBF kernel (a) training set and (b) testing set. The RF model predicted the ECD with an RMSE of 0.23 and R2 of 0.99 in the training set and with an RMSE of 0.42 and R2 of 0.99 in the testing set. The optimum model parameters are shown in Table . Figure a,b are cross-plots of the actual and predicted ECD of the training and testing sets, respectively.
Table 5

Optimum Parameters of the RF Model

parametersoptimum
max_features: [“auto”, “sqrt”, “log2”]Sqrt
max_depth: [3, 4, 5, ..., 30]11
n_estimators: [3, 4, 5, ..., 150]100
Figure 3

Cross-plot of the actual and predicted ECD with RF model (a) training set and (b) testing set.

Cross-plot of the actual and predicted ECD with RF model (a) training set and (b) testing set. The FN model performance with different methods and relationship types is presented in Table . Table shows that the best results were obtained when FNEBM was used with a nonlinear relationship type. The model predicted the ECD with an RMSE of 0.44 and R2 of 0.99 in the training set and with an RMSE of 0.45 and R2 of 0.99 in the testing set. Figure a,b denotes cross-plots of the actual and predicted ECD of the training and testing sets, respectively, when FNEBM with a nonlinear relationship was used.
Table 6

Performance of the FN Model with Different Methods and Relationship Types

FN methodrelationship typeRMSE_trainingR2_trainingRMSE_testingR2_testing
FNFBMnonlinear linear0.360.990.510.98
  0.540.980.550.98
FNEBMnonlinear linear0.440.990.450.99
  0.530.980.550.98
Figure 4

Cross-plot of the actual and predicted ECD with FN model (a) training set and (b) testing set.

Cross-plot of the actual and predicted ECD with FN model (a) training set and (b) testing set. The optimum results of each model are summarized in Table . Table shows that the RF is the most accurate model in estimating the ECD with a low RMSE and high R2 in the training and testing sets. The FN was the second most accurate model, followed by the SVM as the least accurate model.
Table 7

Summary of the Optimum Results of the Models

modelRMSE_trainingR2_trainingRMSE_testingR2_testing
RF0.230.990.420.99
FN0.440.990.450.99
SVM0.540.970.580.97

Validation of the Developed RF

The best model, RF, was selected to be validated using a total of 1152 unseen data points collected from Well-2. The compatibility of the input parameters in Well-2 was checked to ensure they are in the same range as the data set, which was used to train the RF model. Table shows the statistical parameters of the data set of Well-2. The RF predicted the ECD in Well-2 with an RMSE of 0.35 and an R2 of 0.95. Figure shows the actual and predicted ECD as a function of the depth index of Well-2. The depth index was used to hide the actual depth of the drilled section.
Table 8

Statistical Parameters of the Data set of Well-2 (1152 Data Points)

statistical parametersQ (gal/min)HL (klbf)ROP (ft/h)RS (RPM)SPP (psi)WOB (klbf)T (kft.lbf)ECD (Pcf)
minimum273.6265.53.975.22865.45.64.287.7
maximum288.0283.857.2136.03449.419.57.493.4
mean279.6275.029.8112.03099.914.15.790.6
standard deviation2.44.78.97.9123.92.40.71.6
skewness0.55–0.040.00–0.840.25–0.640.230.13
kurtosis1.14–1.44–0.102.52–1.070.41–1.02–1.41
Figure 5

Actual and predicted ECD as a function of the depth index of Well-2.

Actual and predicted ECD as a function of the depth index of Well-2.

PCA for Dimensionality Reduction

The input variables of the original data set collected from Well-1 were standardized using eq . Then, the data set was transformed to PCA space using Python library’s Scikit-Learn. Table shows a sample of the training set after transformation to PCA space.
Table 9

Sample of the Training Set after Transformation to PCA Space

samplePC1PC2PC3PC4PC5PC6PC7ECD (Pcf)
10.01575.6083–0.9000–1.16460.5203–0.5557–0.594683.57
2–0.10045.27290.1541–1.56671.3405–0.0313–0.109083.45
30.20064.9792–0.1247–1.51611.41030.56790.176283.42
40.41354.8381–0.1609–1.30591.45390.78500.221283.44
5–0.32425.09520.0931–1.21101.4730–0.04220.081683.47
It is important to study the variation that each PC accounts for in the data set to perform dimensionality reduction. Figure is a screen plot, which is a graphical representation of the percentages of the variation that each PC accounts for in the whole data set. The first thing to notice is that the first principal component PC1 accounts for 36.42% of the variation in the data set; the second principal component PC2 accounts for 26.68%; the third principal component PC3 accounts for 17.30%; the fourth principal component PC4 accounts for 13.01%; the fifth principal component PC5 accounts for 4.47%; the sixth principal component PC6 accounts for 1.29%, and the seventh principal component PC7 accounts for 0.75%. This means that PC1, PC2, PC3, and PC4 directions collectively explain 93.41% of the total variation in the data set, while PC5, PC6, and PC7 combined explain only 6.51% of the total variation in the data set. Thus, even though the points in the data set form a cloud in a dimensional space (R7), PCA shows that these points cluster near a four-dimensional plane (R4) spanned by PC1, PC2, PC3, and PC4.
Figure 6

Percentages of the variation that each PC accounts for in the whole data set.

Percentages of the variation that each PC accounts for in the whole data set. The contribution of each variable (Q, HL, ROP, RS, SPP, WOB, or T) in each principal component is presented in Table . How all of this needs to be interpreted? For example, in studying PC1, the second entry “0.58”, HL, is the largest, which means a change in one unit of HL tends to affect the ECD more than a change of one unit of Q, ROP, RS, SPP, WOB, or T. The first entry “0.52”, which corresponds to Q, is the next most important factor in determining the ECD. On the other hand, the second last entry “0.14”, WOB, is the least important in determining the ECD. Similarly, other PCs can be interpreted.
Table 10

Contribution of Variables in Each Principal Component

variablePC1PC2PC3PC4PC5PC6PC7
Q0.520.270.100.270.480.030.58
HL0.580.030.270.000.120.470.59
ROP0.190.300.080.890.150.230.01
RS0.260.530.310.110.730.100.01
SPP0.250.610.020.300.350.340.48
WOB0.140.140.850.050.260.370.17
T0.450.410.290.150.090.680.23
The PCA-based RF model performance in predicting the ECD in the same testing set used in Section , while using different numbers of PCs, was evaluated and compared with the SVM and FN, which were trained and tested using 100% variation of the data set (i.e., seven dimensions) as discussed in Section . Table shows that as the number of PCs for the data set increases from one principal component to four principal components, the R2 and RMSE improve. However, when more than four principal components were considered, RMSE increased. The PCA-based RF with only four principal components outperformed the SVM, with an RMSE of 0.54 and R2 of 0.97. In addition, the PCA-based RF with only four principal components performed almost similar to the FN that had an RMSE of 0.45 and R2 of 0.99. Furthermore, even if only two dimensions PC1 and PC2 (63.10% of variation) were considered, the PCA-based RF would perform almost similar to the SVM with an RMSE of 0.63 and R2 of 0.96. In other words, the PCA-based RF model with four inputs (i.e., PC1, PC2, PC3, and PC4) performed better than the SVM and almost similar to the FN, which were trained with seven inputs (i.e. Q, HL, ROP, RS, SPP, WOB, and T). This showed how the PCA technique is powerful enough to transform the data set from a higher dimensional space to a lower dimensional space without loss of any important information, which in turn increases the speed of the training process for a model.
Table 11

Comparison of PCA-Based RF Performance with FNN and SVM Using Different PCs

  PCA-RF testing
FN testing
SVM testing
no. PCvariation %RMSER2RMSER2RMSER2
PC136.421.780.700.450.990.580.97
PC1 and PC263.100.630.96    
PC1 to PC380.400.620.96    
PC1 to PC493.410.540.97    
PC1 to PC597.880.570.97    
PC1 to PC699.170.590.97    
PC1 to PC71000.600.97    

Conclusions

Nowadays, modern drilling involves multiple interconnecting activities. Therefore, obtaining real-time information about ongoing operations is crucial for safe and efficient drilling operations. Three intelligent machines SVM, RF, and FN were developed to predict the ECD while drilling horizontal wells. PCA was applied to reduce the data set dimensions and compare the result of the PCA-based RF with the SVM and FN. Based on the results, the following can be concluded: The RF achieved the lowest RMSE and the highest R2 compared to the FN and SVM. The RF predicted the ECD with an RMSE of 0.23 and R2 of 0.99 in the training set and with an RMSE of 0.42 and R2 of 0.98 in the testing set. The FN achieved the second-lowest RMSE of 0.44 and 0.45 in the training and testing sets, respectively. The R2 was as high as the RF of 0.99 in the training and testing sets. The SVM with the kernel type had the highest RMSE and the lowest R2 when compared with that of the RF and FN models. The SVM with the optimum parameters predicted the ECD with an RMSE of 0.54 and R2 of 0.97 in the training set, while the RMSE and R2 were 0.58 and 0.97, respectively, in the testing set. The RF predicted the ECD in Well-2 with an RMSE of 0.35 and R2 of 0.95. The PCA-based RF model with only four principal components outperformed the SVM with an RMSE of 0.54 and R2 of 0.97.
  1 in total

1.  Real-Time Prediction of Rheological Properties of Invert Emulsion Mud Using Adaptive Neuro-Fuzzy Inference System.

Authors:  Ahmed Alsabaa; Hany Gamal; Salaheldin Elkatatny; Abdulazeez Abdulraheem
Journal:  Sensors (Basel)       Date:  2020-03-17       Impact factor: 3.576

  1 in total
  1 in total

1.  Real-time prediction of formation pressure gradient while drilling.

Authors:  Ahmed Abdelaal; Salaheldin Elkatatny; Abdulazeez Abdulraheem
Journal:  Sci Rep       Date:  2022-07-05       Impact factor: 4.996

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.