| Literature DB >> 33564733 |
Zeeshan Tariq1, Mohamed Mahmoud1, Mohamed Abouelresh1, Abdulazeez Abdulraheem1.
Abstract
Prediction of thermal maturity index parameters in organic shales plays a critical role in defining the hydrocarbon prospect and proper economic evaluation of the field. Hydrocarbon potential in shales is evaluated using the percentage of organic indices such as total organic carbon (TOC), thermal maturity temperature, source potentials, and hydrogen and oxygen indices. Direct measurement of these parameters in the laboratory is the most accurate way to obtain a representative value, but, at the same time, it is very expensive. In the absence of such facilities, other approaches such as analytical solutions and empirical correlations are used to estimate the organic indices in shale. The objective of this study is to develop data-driven machine learning-based models to predict continuous profiles of geochemical logs of organic shale formation. The machine learning models are trained using the petrophysical wireline logs as input and the corresponding laboratory-measured core data as a target for Barnett shale formations. More than 400 log data and the corresponding core data were collected for this purpose. The petrophysical wireline logs are γ-ray, bulk density, neutron porosity, sonic transient time, spontaneous potential, and shallow resistivity logs. The corresponding core data includes the experimental results from the Rock-Eval pyrolysis and Leco TOC measurements. A backpropagation artificial neural network coupled with a particle swarm optimization algorithm was used in this work. In addition to the development of optimized PSO-ANN models, explicit empirical correlations are also extracted from the fine-tuned weights and biases of the optimized models. The proposed models work with a higher accuracy within the range of the data set on which the models are trained. The proposed models can give real-time quantification of the organic matter maturity that can be linked with the real-time drilling operations and help identify the hotspots of mature organic matter in the drilled section.Entities:
Year: 2020 PMID: 33564733 PMCID: PMC7864083 DOI: 10.1021/acsomega.0c03751
Source DB: PubMed Journal: ACS Omega ISSN: 2470-1343
TOC and Tmax Range Describing the Level of Maturity
| parameter | no hydrocarbon | maturity range | postmaturity range | refs |
|---|---|---|---|---|
| TOC, wt % | less than 0.5 | 0.5–2 | greater than 2 | ( |
| less than 435 | 435–465 | greater than 465 | ( |
Summary of the Research Related to the Prediction of Organic Matter in Shale
| refs | study conducted | technique | method type | input parameters | geological field study |
|---|---|---|---|---|---|
| Tan et al.[ | prediction of TOC | artificial intelligence | epilson-SVR, nu-SCR, SMO-SVR, and RBF | CNL, GR, AC, K, TH, U, PE, RHOB, and RT | Huangping syncline, China |
| Rui et al.[ | prediction of TOC | artificial intelligence | SVM | wireline log data such as RHOB, GR, SP, RT, and DT | Beibu Gulf basin |
| Lawal et al.[ | prediction of TOC | artificial intelligence | ANN | XRD data: SiO2, Al2O3, MgO, and CaO | Devonian Shale |
| Sultan[ | prediction of TOC | artificial intelligence | self-adaptive differential evolution-based ANN | well logs: GR, DT, RT, and RHOB | Devonian Shale |
| Mahmoud[ | prediction of TOC | artificial intelligence | ANN | well logs: GR, DT, RT, and RHOB | Devonian Shale |
| Zhao et al.[ | prediction of TOC | regression | nonlinear | CNL | Ordos Basin in China and Bakken Shale of North Dakota |
| Wang et al.[ | prediction of TOC | regression | nonlinear | DT and RT | Sichuan Basin, Southern China |
| Alizadeh et al.[ | prediction of TOC and | artificial intelligence | ANN | DT and RT | Dezful Embayment, Iran |
| Handhal et al.[ | prediction of TOC | artificial intelligence | SVR, ANN, KNN, random forest, and rotation forest | GR, RHOB, NPHI, RllD, and DT | Rumaila Oil Field, Iran |
| Wang et al.[ | TOC, | artificial intelligence | ANN | RHOB, NPHI, RT, and DT | Bohai Bay Basin, China |
GR = γ-ray, RHOB = bulk density, LLD = deep lateral log, LLS = shallow lateral log, MSFL = microspherical focused log, RILD = deep induction resistivity log, DT = compressional wave travel time, TH = thorium, U = uranium, K = potassium, RT = resistivity log, NPHI = neutron porosity, SP = spontaneous potential, CNL = compensated neutron log, PE = photoelectric index, SiO2 = silicondioxide, Al2O3 = aluminiumdioxide, MgO = magnesium oxide, and CaO = calcium oxide.
Step-by-Step Pseudocode for the Proposed PSO-ANN Algorithm for Thermal Maturity Parameter Prediction
| steps | working |
|---|---|
| 1 | start |
| 2 | set input variables |
| 3 | initialize parameters of ANN such as learning rate, activation functions, etc. |
| 4 | vary the number of hidden layers (sensitivity of hidden layers, 1–3) |
| 5 | vary the number of neurons in the hidden layer (sensitivity of neurons, 5–30) |
| 5 | select the learning algorithm of ANN |
| 6 | select the learning rate [0, 1] for the selected learning algorithm |
| 7 | train and test the ANN model and |
| 8 | evaluate the objective function for a minimum convergence value |
| 9 | extract weights and biases from the trained model |
| 10 | initialize parameters of PSO algorithm such as the number of iterations, population of particles, cognitive and social accelerations, and initial and final inertia weights |
| 11 | set range for sample search space of each extracted weights and biases |
| 12 | feed extracted weights and biases in a PSO algorithm as the initial population |
| 13 | evaluate the objective function for a minimum convergence value |
| 14 | run the iterative process until the stopping criterion |
| 15 | pick the global best solution |
| 16 | set optimum weights and biases from the globally best model in the network for the prediction of thermal maturity parameters |
| 17 | end |
stopping criterion = a maximum number of iterations are attained or a maximum level of inactivity is reached.
Figure 1Workflow chart of the proposed PSO-ANN algorithm to obtain thermal maturity parameters.
Figure 2Optimal number of neurons in the middle layer using an objective function.
Figure 3General topography of the proposed ANN model for the prediction of thermal maturity parameters.
Optimum Values for the Proposed ANN Models
| parameters of the ANN model | range | TOC | ||||
|---|---|---|---|---|---|---|
| number of input parameters | 6 | 6 | 6 | 6 | 6 | 6 |
| middle layer(s) | 1–3 | 1 | 1 | 1 | 1 | 1 |
| neurons in the middle layer | 5–15 | 10 | 10 | 10 | 10 | 10 |
| learning algorithm | quasi-newton, conjugate gradient, Levenberg–Marquardt (LM), Newton’s method, gradient descent (GD), resilient backpropagation (RB), Fletcher–Powell conjugate gradient, one-step secant | LM | RB | RB | LM | GD |
| rate of learning, α | 0.1–0.5 | 0.15 | 0.2 | 0.10 | 0.16 | 0.18 |
| middle-layer transfer function | tangential sigmoidal (tansig), logarithmic sigmoidal, hyperbolic sigmoidal, linear, rectified linear unit | tansig | tansig | tansig | tansig | tansig |
| epochs | 100–500 | 150 | 115 | 125 | 180 | 260 |
| outer-layer transfer function | linear | linear | linear | linear | linear | linear |
Figure 4Training and testing of the Tmax model.
Figure 5Training and testing of the S1 model.
Figure 6Training and testing of the S2 model.
Figure 7Training and testing of the S3 model.
Figure 8Training and testing of the TOC model.
Figure 9Comparison of conventional ANN and PSO-ANN in terms of the coefficient of determination on an overall data set.
Figure 10Diagram showing the fractions of the total organic matter.
Ranges of the Data Used for AI Modeling
| GR,
API | RHOB (g/cc) | NPHI (vol/vol) | Δ | SP
(mV) | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| models | min | max | min | max | min | max | min | max | min | max | min | Max |
| 18.229 | 417.06 | 2.37 | 2.86 | 0.003 | 0.33 | 4.8 | 1283.34 | 45.27 | 93.48 | –154.18 | –28.62 | |
| 19.120 | 336.73 | 2.37 | 2.86 | 0.003 | 0.33 | 1.3 | 1027.90 | 45.27 | 93.39 | –154.18 | –28.62 | |
| 18.229 | 372.29 | 2.37 | 2.86 | 0.003 | 0.33 | 1.3 | 1283.34 | 45.27 | 93.39 | –154.18 | –28.81 | |
| 18.229 | 417.06 | 2.37 | 2.83 | 0.003 | 0.29 | 4.8 | 1088.66 | 45.27 | 92.08 | –154.18 | –28.81 | |
| TOC (wt %) | 55.500 | 359.76 | 2.20 | 2.68 | 0.03 | 0.33 | 6.0 | 148.87 | 62.05 | 93.39 | –86.69 | –29.25 |
Figure 11Well logs’ input data (AT90 is a RILD log).
Figure 12Frequency distribution of Tmax, S1, S2, S3, and TOC.
Figure 13Relative importance of the input parameters such as GR, ρ, NPHI, AT90 (RILD), Δtc, and SP logs with the output parameters such as Tmax, S1, S2, S3, and TOC.
Statistical Indicators of Model Performance Evaluationa
| goodness-of-fit test | mathematical expression |
|---|---|
| average absolute percentage error | |
| mean absolute error | |
| root-mean-square error | |
| coefficient of determination |
Ymeasured is the measured value of TOC, Ypredicted is the estimated value from the model, and n is the total number of samples.
Weights and Biases of the Proposed Model for Tmax Prediction
| weights
between input and hidden layers ( | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| hidden layer neurons ( | GR | ρ | NPHI | Δ | SP | weights between hidden and output layers ( | hidden layer bias ( | output layer bias ( | |
| 1 | –1.7941 | 0.2967 | 0.2405 | 3.9757 | –0.6821 | 0.1778 | –1.7717 | 5.1003 | 1.4308 |
| 2 | –2.0779 | 0.8055 | 1.2726 | –0.5317 | –0.2028 | –0.3555 | 2.9596 | 1.0329 | |
| 3 | –0.2541 | –1.2323 | 2.3337 | –0.0487 | 4.0613 | 1.0812 | 0.5933 | 6.0397 | |
| 4 | 0.1803 | –1.5660 | –0.7003 | 1.9781 | 0.3196 | 0.4630 | –1.3631 | 2.3029 | |
| 5 | –1.8247 | 0.7878 | 2.1122 | 1.7892 | 0.2356 | 2.8205 | 1.0260 | –2.4688 | |
| 6 | 0.8137 | –1.3633 | –0.8132 | 1.1902 | –0.4400 | 0.0050 | 2.5527 | 1.4510 | |
| 7 | –0.4772 | –3.1851 | –0.8343 | 0.4839 | 0.3739 | –3.1749 | –0.5414 | 0.1539 | |
| 8 | –1.1043 | 3.5500 | –3.5252 | –0.9498 | 2.1798 | 4.9600 | –0.1921 | –3.1789 | |
| 9 | –1.7033 | 0.4957 | –0.6426 | –0.4278 | –1.3891 | 0.1037 | 1.8245 | –4.7481 | |
| 10 | 2.3884 | –2.4291 | 1.0258 | –0.2708 | –4.4849 | 2.0855 | –0.4637 | 1.1179 | |
Weights and Biases of the proposed Model for S1 Prediction
| weights
between input and hidden layers ( | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| hidden layer neurons ( | GR | ρ | NPHI | Δ | SP | weights between hidden and output layers ( | hidden layer bias ( | output layer bias ( | |
| 1 | –0.0872 | 2.3795 | 0.61 | 0.9346 | –1.0302 | 2.6013 | 2.2032 | 2.7471 | –1.125 |
| 2 | 2.9553 | 1.196 | 0.9532 | –0.1497 | 2.0403 | –0.6852 | –1.5668 | 3.5566 | |
| 3 | 0.9799 | –2.0489 | –0.5193 | 1.0599 | 1.1969 | 3.2706 | –2.0522 | –0.5307 | |
| 4 | –0.3019 | 1.5661 | 0.1572 | 0.0954 | 0.0242 | –2.0559 | –3.1289 | 1.1359 | |
| 5 | 1.8638 | –0.8322 | –3.3966 | 0.6946 | 2.8564 | 3.4534 | –1.377 | –3.9418 | |
| 6 | –3.2506 | 0.8794 | 1.5293 | 1.2903 | 0.8682 | 0.4842 | –3.4481 | –0.2281 | |
| 7 | 1.7284 | –1.3916 | 4.3578 | –2.3109 | –5.3719 | 0.012 | –1.0532 | 0.6863 | |
| 8 | –2.3325 | 0.5383 | 1.3951 | –0.8393 | 0.687 | 0.3473 | 3.2165 | –2.2594 | |
| 9 | 0.6438 | 1.1076 | 3.1629 | 0.6633 | 0.8062 | –0.8584 | 1.4581 | 3.789 | |
| 10 | 1.7403 | –1.5715 | 1.5465 | –2.1928 | –0.0194 | –2.249 | –0.7571 | 1.6332 | |
Weights and Biases of the Proposed Model for S2 Prediction
| weights
between input and hidden layers ( | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| hidden layer neurons ( | GR | ρ | NPHI | Δ | SP | weights between hidden and output layers ( | hidden layer bias ( | output layer bias ( | |
| 1 | –3.615 | –3.658 | 5.426 | 1.372 | 2.019 | –5.133 | –0.406 | –2.487 | 0.711 |
| 2 | 5.917 | –0.778 | 1.821 | 1.348 | –2.140 | 1.007 | –0.509 | 5.152 | |
| 3 | –2.575 | –0.020 | 1.776 | –0.650 | –1.119 | 1.817 | 5.629 | –0.469 | |
| 4 | –0.339 | 0.970 | –1.707 | –1.300 | –0.551 | –0.881 | 1.030 | –0.459 | |
| 5 | –0.890 | –0.478 | 0.944 | –9.386 | –0.114 | 2.103 | 2.762 | –11.697 | |
| 6 | –1.869 | 0.389 | 0.903 | 1.473 | 0.715 | –0.465 | 2.594 | 1.915 | |
| 7 | 1.196 | –9.027 | –3.565 | 1.744 | –4.834 | –6.791 | –0.331 | 1.166 | |
| 8 | –2.109 | 1.027 | 0.875 | 0.538 | 1.649 | –0.208 | –2.602 | 0.392 | |
| 9 | –1.703 | –0.176 | 1.199 | –0.617 | –1.307 | 1.459 | –6.972 | –0.500 | |
| 10 | 4.054 | –6.402 | –1.309 | –5.677 | –6.265 | 17.736 | 0.621 | –10.018 | |
Weights and Biases of the Proposed Model for S3 Prediction
| weights
between input and hidden layers ( | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| hidden layer neurons ( | GR | ρ | NPHI | Δ | SP | weights between hidden and output layers ( | hidden layer bias ( | output layer bias ( | |
| 1 | –1.366 | –0.203 | 0.630 | –0.508 | –1.215 | –1.586 | –2.372 | 2.103 | –2.081 |
| 2 | 0.248 | 0.581 | –0.757 | –4.469 | 0.632 | 1.145 | 3.125 | –3.801 | |
| 3 | –1.266 | 3.104 | 2.077 | –0.856 | –1.393 | –2.028 | 0.958 | 1.611 | |
| 4 | –0.034 | –0.877 | 0.945 | –4.088 | –0.183 | 0.358 | 2.039 | –3.462 | |
| 5 | –0.485 | 0.079 | 2.798 | –0.098 | –2.098 | –2.592 | 1.451 | –1.440 | |
| 6 | –0.131 | 1.104 | –0.111 | 0.917 | –1.104 | –0.280 | 3.982 | 0.190 | |
| 7 | 0.006 | 0.473 | 0.145 | –3.503 | –0.081 | 0.578 | –3.989 | –3.248 | |
| 8 | 0.342 | 0.235 | –0.188 | –0.969 | –1.883 | 0.090 | –2.849 | –1.548 | |
| 9 | 0.459 | –0.080 | 1.096 | –1.149 | –1.765 | 2.086 | 2.723 | 1.431 | |
| 10 | –1.719 | –0.903 | –2.539 | 0.334 | –3.960 | 0.820 | 0.985 | 4.271 | |
Weights and Biases of the Proposed Model for TOC Prediction
| weights
between input and hidden layers ( | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| hidden layer neurons ( | GR | ρ | NPHI | Δ | SP | weights between hidden and output layers (w2) | hidden layer bias ( | output layer bias (b2) | |
| 1 | 1.759 | –1.040 | 1.233 | 2.680 | –0.320 | –1.225 | 1.372 | –1.683 | 1.923 |
| 2 | 0.875 | 0.288 | 0.309 | –1.399 | –0.701 | 0.682 | 4.431 | –0.756 | |
| 3 | 0.990 | –0.414 | –1.286 | –1.252 | –0.283 | 3.069 | –1.250 | –1.101 | |
| 4 | –1.100 | –1.601 | –1.629 | –3.078 | 0.297 | –0.882 | 1.538 | –3.488 | |
| 5 | 2.962 | –1.130 | 0.781 | –2.217 | –1.349 | 0.861 | –0.787 | 0.840 | |
| 6 | –0.886 | 1.227 | 0.386 | 6.639 | 2.619 | –3.780 | 1.059 | 2.219 | |
| 7 | 0.417 | –1.863 | 0.815 | –0.082 | 1.125 | –1.183 | –1.524 | 2.046 | |
| 8 | –0.847 | –1.005 | 0.479 | 0.991 | 0.410 | 0.273 | –0.556 | –2.068 | |
| 9 | –0.606 | –1.045 | –1.203 | 0.960 | 1.449 | 0.264 | 2.000 | 1.132 | |
| 10 | 2.624 | –2.675 | –0.025 | –0.704 | 0.800 | –1.333 | 1.013 | 3.426 | |