Next Article in Journal
Forming Ni-Fe and Co-Fe Bimetallic Structures on SrTiO3-Based SOFC Anode Candidates
Previous Article in Journal
Development of a Design Tool for Performance Estimation and Validation Proton Exchange Membrane Fuel Cell: Verification and Validation for 20 KW Commercial Fuel Cell
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Analysis of the Prediction of Gas Condensate Dew Point Pressure Using Advanced Machine Learning Algorithms †

1
Department of Petroleum Engineering, University of Houston, Houston, TX 77023, USA
2
Shell International E&P Inc., Westhollow Technology Center, 3333 Hwy 6 South, Houston, TX 77082, USA
3
Xecta Digital Labs, Houston, TX 77201, USA
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper presented in 2021 SPE Annual Technical Conference and Exhibition, Dubai, United Arab Emirates, 15 September 2021 (SPE-205997-MS).
Current address: Citibank, 3800 Citibank Center, Tampa, FL 33610, USA.
Fuels 2024, 5(3), 548-563; https://doi.org/10.3390/fuels5030030
Submission received: 8 May 2024 / Revised: 19 June 2024 / Accepted: 28 June 2024 / Published: 16 September 2024

Abstract

:
Dew point pressure (DPP) emerges as a pivotal factor crucial for forecasting reservoir dynamics regarding condensate-to-gas ratio and addressing production/completion hurdles, alongside calibrating EOS models for integrated simulation. However, DPP presents challenges in terms of predictability. Acknowledging these complexities, we introduce a state-of-the-art approach for DPP estimation utilizing advanced machine learning (ML) techniques. Our methodology is juxtaposed against published empirical correlation-based methods on two datasets with limited sizes and diverse inputs. With superior performance over correlation-based estimators, our ML approach demonstrates adaptability and resilience even with restricted training datasets, spanning various fluid classifications. We acquired condensate PVT data from publicly available sources and GeoMark RFDBASE, encompassing dew point pressure (the target variable), as well as compositional data (mole percentages of each component), temperature, molecular weight (MW), and specific gravity (SG) of heptane plus, which served as input variables. Before initiating the study, thorough assessments of measurement quality and results using statistical methods were conducted leveraging domain expertise. Subsequently, advanced ML techniques were employed to train predictive models with cross-validation to mitigate overfitting to the limited datasets. Our models were juxtaposed against the foremost published DDP estimators utilizing empirical correlation-based methods, with correlation-based estimators also trained on the underlying datasets for equitable comparison. To improve outcomes, pseudo-critical properties and artificial proxy features were utilized, leveraging generalized input data.

1. Introduction

For gas condensates, dew point pressure is harder to calculate or predict due to the complexity of the behavior of the heavier components. In addition, measurement of the dew point pressure is also more difficult with respect to the bubble-point pressure as it often involves visual justification for the measurements. Current correlations published in the literature have limitations both in their predictive capabilities and as well as their range of validities. Therefore, traditional direct experimental determination is still essential for the identification of dew points. The first published dew point pressure correlation can be dated back to 1947 by Sage and Olds [1]. Sage and Olds [1] developed a simple correlation that uses stock tank oil API gravity, gas–oil ratio (GOR), and temperature to predict dew point pressure. Nemeth and Kennedy [2] developed a more complex correlation with 11 coefficients to calculate dew point pressure using compositional information, the bulk properties of the plus fraction (molecular weight of heptane plus and specific gravity of heptane plus), and reservoir temperature. Although the Nemeth and Kennedy [2] correlation was established about five decades ago, it is still one of the most widely used empirical correlations for quick calculation of dew point pressure. Elsharkawy [3] proposed a new correlation for dew point pressure prediction, which uses the same input parameters as Nemeth and Kennedy [2], but the resulting equation is more complex with higher degrees of freedom (19 coefficients in total). Apart from the traditionally developed empirical correlations for dew point pressure, Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5] used gene expression programming or genetic programming (GP or GEP) to develop a new set of correlation equations for dew point pressure. Such genetic programming (GP) and gene expression programming (GEP) methods are advantageous because they produce an explicit equation similar to traditional empirical correlations. However, the equations derived from these methods tend to be more complex, often containing numerous terms and significant nonlinearity. This complexity can sometimes lead to overfitting, where the model captures noise in the training data rather than the underlying trend. Therefore, more flexible machine learning models were developed for dew point pressure prediction [6,7]. Jalali et al. [6] used an artificial neural network (ANN) as the machine learning algorithm and utilized compositional data, molecular weight of heptane plus ( M W C 7 + ), and temperature as predictors on a 111-sample dataset. Alarfaj [8] compared several machine learning models including ANN, support vector machine (SVM), and basic decision trees using the same predictors as Jalali et al. [6]. The dataset of Alarfaj [8] includes the results from 98 PVT reports. Although the dataset size needed for machine learning depends on the complexity of the problem and the complexity of learning algorithms, a dataset with about 100 samples is still a typical small dataset [9]. Neural networks often exhibit severe overfitting problems, particularly when trained over a large number of iterations, especially with small datasets [10]. Alzahabi [7] compared the performance of linear regression, random forest, generalized additive models (GAMs with local smoothing), and neural networks (with a single hidden layer) on a dataset of 667 samples. The study concluded that these four models performed similarly in terms of accuracy. However, it is important to note that while neural networks can be powerful, they require careful tuning and sufficient data to avoid overfitting and achieve optimal performance.
The advanced machine learning algorithm chosen for this study is extreme gradient boosting, commonly abbreviated as XGBoost [11,12,13], due to its robustness and strong ability to capture nonlinear behavior, making it highly predictive. XGBoost [12,13], developed by Chen and Guestrin [13], quickly emerged as one of the most powerful and popular advanced machine learning algorithms among data scientists.
Based on the gradient boosting algorithm pioneered by Friedman [14,15], XGBoost utilizes an ensemble of weak prediction trees for robust prediction (see Sinha et al. [16] and Sinha et al. [17] to read the details about the mathematics gradient boosting methods (GBMs)). Its key advantages over other gradient boosting methods include the following features:
(1)
The incorporation of regularization terms (γ and λ in Figure 1a) for leaf weights ( w j in Figure 1a) and the number of leaf nodes of trees (T), restricting their values during loss minimization; this reduces the influence of each individual tree, allowing room for future trees to improve the model.
(2)
The use of second-order partial derivatives (as shown in Figure 1b) as an approximation of the loss function, providing more information about the gradient direction and the path to minimize the loss function.
(3)
High prediction capabilities facilitated by pre-sorted and histogram-based algorithms for splitting at each node; XGBoost [12,13] explores various combinations of splitting and selects the one with the best value of the scoring function derived from minimizing the loss function XGBoost [12,13].
(4)
Efficient handling of sparse data and missing values in inputs.
Furthermore, to enhance the prediction of dew point pressure and develop a robust model, gas pseudocritical properties (calculated from correlations) and artificial proxy features were incorporated into XGBoost [12]. This correlation aims to further improve the predictive performance of the model.

2. Dataset Analysis

In this study, the gas condensate PVT data were obtained from GeoMark RFDBASE (RFDbase—Rock and Fluid Database by GeoMark Research, Ltd. [18]), which contains company data as well as general access data. In order to research the ability of the advanced machine learning algorithms to predict dew point pressure, we used the entire mix of PVT data with no upper and lower range limits and with no region constraints. Based on the previous research on predicting dew point pressure [2,3,6,7,8,19,20], mole percentages of methane through heptane plus, nitrogen, carbon dioxide and hydrogen sulfide, reservoir temperature, molecular weight of reservoir fluid, and molecular weight and specific gravity of heptane plus are widely used as predictors. Thus, these predictors were also considered during our data cleaning and categorization process before the application of machine learning methods. In addition to staying aligned with what became almost the norm in terms of the selected predictors in the literature, we did not want to use other input data as potential add-on parameters in addition to these indicators as they tend to be hard to find/measure and also expensive, which almost defeat the purpose of estimation of the dew point pressure. Generally, there is always a fine balance between the required input parameters and the predicted quantities, and this case is no different.
We investigated the available data based on the availability and validity of the input proxies unless there were other known issues. The raw dataset we obtained had 521 gas samples with relevant PVT information. Among those, there were 96 dry gas samples that we discarded as they did not have dew point pressure measurements, leaving 425 relevant samples in the dataset. Second, we used compositional data (mole % of each component) as predictors so we needed to exclude samples with missing or invalid compositional information. After eliminating 4 of the samples that did not have usable compositional information, the number of samples in the set was reduced to 421. We also observed that 2 samples had missing molecular weights in the database, reducing the set to 419 samples. In addition, 53 samples did not have any recorded specific gravity values of the plus fractions ( S G C n + ) along with the 2 samples identified with an incorrect molecular weight of heptane plus on the database. Therefore, there were 364 samples left in the final checked dataset. Furthermore, reservoir temperature was not provided for 1 sample, 1 sample had extremely high oil-based mud contamination, and 20 samples did not have accurate plus fluid properties leading to the final 342 samples in the dataset.
The final dataset of 342 samples was used with 16 predictors, namely the measured mole % of N 2 , C O 2 , H 2 S , C 1 , C 2 , C 3 , i C 4 , n C 4 , i C 5 , n C 5 , C 6 , and C 7 + ; molecular weight of C 7 + ; reservoir temperature; and specific gravity of C 7 + . The statistical information of the corresponding final dataset is shown in Table 1.

3. Development of the New Model

Several alternative advanced machine learning algorithms are included in Appendix A, along with their performance dimensions. Based on our study using the dataset, the results from XGBoost [12,13] were found to be the best among those compared in Appendix A. Therefore, in this section, the outcomes and details of the XGBoost [12,13] model are discussed.
XGBoost
Traditionally, researchers have used a portion of this dataset to train the advanced machine learning model and leave the rest of the dataset as test data [2,3,6,8,19,20]. However, for the dew point pressure prediction models that have been published, the datasets were usually under 500 samples, and the test dataset was usually under 100 samples [6,8,19,20]. In this case, the test dataset cannot cover the whole range of dew point pressure and is even chosen randomly, so the errors may vary depending on the random seed. In this study, K-fold cross-validation was applied to make full use of our dataset and thus show a fair comparison between the advanced machine learning algorithm and empirical correlations. For K-fold cross-validation, the whole dataset was randomly split into K subsets of the same size. For each fold or subset, the rest of the dataset was used to train the model and then predict this subset of data. For example, if we were using 5-fold cross-validation, in each iteration, 4-fold data were used as training data, and 1-fold data were used as test data until all data samples were predicted. The merit of K-fold cross-validation is that each sample in the dataset can be predicted by the model building without this sample itself. The usual choice of K-fold cross-validation is K = 5 or K = 10 [21]. Additionally, 10-fold cross-validation was applied to the dataset in order to maintain a reasonable training dataset size.
In addition to the primary input variables, we also used artificial physical proxy features, including groups of input parameters, such as pseudocritical properties, via a selected correlation (Sutton [22]).
Sutton’s correlations
Pseudocritical properties of gases were applied to adjust the performance of XGBoost [12,13]. Gas mixture pseudocritical properties can be estimated from Sutton’s correlations based on gas-specific gravity (Sutton [22]).
Sutton [22] correlations were determined based on over 3200 databases of associated gas compositions and suitable for gas condensates, expressed as follows:
T p c = 164.3 + 357.7 γ g H C 67.7 γ g H C 2
P p c = 744 125.4 γ g H C + 5.9 γ g H C 2
where
γ g H C = M W   o f   t h e   H y d r o c a r b o n   g a s   ( H C )   g a s 28.967
Artificial proxy features (APFs)
Artificial proxy features were combined with XGBoost [12,13] and Sutton’s correlations [22] for gas pseudocritical properties to develop a physical sound and to improve the accuracy of the model predictions. This is required to suitably model the effects of all variables corresponding to dew point pressure. Each artificial proxy feature was selected based on feature correlations with dew point pressure as a target, and a high-feature correlation distribution was observed, as shown in Figure 2 and Table 2. This indicates the following:
(a)
The molecular weight of heptane plus fraction is the most important for correlation in controlling the dew point pressure. Considering this observation, we initially attempted to develop two artificial proxy features, which are explained as follows:
  • The first artificial proxy feature (APF#1) was created by using the product of molecular weight and the mole percent of heptane plus fraction.
    A P F # 1 = M W C 7 +   z C 7 + 3
  • The mole percent of methane was combined with the molecular weight of heptane plus fraction to achieve the second artificial proxy feature (APF#2).
    A P F # 2 = M W C 7 +   z C 1
(b)
Mixture pseudocritical properties ( T p c and P p c ) contribute more to the dew point pressure than reservoir temperature and mole percentages of light, intermediate, and non-hydrocarbon components. The corrections of gas pseudocritical properties were carried out by adjusting the pseudocritical properties of heptane plus using Matthews [23] correlations as the third and fourth artificial proxy features (APF#3 and APF#4).
A P F # 3 = T p c c o r r = E X P T c C 7 +   /   T p c
A P F # 4 = P p c c o r r = E X P P c C 7 + /   P p c
(c)
The effect of reservoir temperature was also needed in a different functional form. Therefore, three artificial proxy features were defined by integrating reservoir temperature with pseudocritical temperature ( T p c ).
A P F # 5 = T R /   T p c 2
A P F # 6 = E X P T R /   T p c
A P F # 7 = T p c / T R
(d)
The mole percent of intermediate ( i C 4 , n C 4 , i C 5 , n C 5 , and C 6 ) and H 2 S has different effects on the dew point pressure, as their interaction with the heptane plus fraction differs from the more volatile segment of the fluid. The last artificial proxy feature (APF#8) was defined using the subject’s intermediate fraction and H2S fraction.
A P F # 8 = z C 7 + z i C 4 + z n C 4 + z i C 5 + z n C 5 + z C 6 + z H 2 S

4. Results and Discussion

Dew point pressure correlations determined by Nemeth and Kennedy [2], Elsharkawy [3], Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5] were used for comparison against our new model.
To be able to perform an objective comparison against the selected set of leading dew point pressure correlations, the models were re-trained using the same selected dataset for this study. The objective function setup for tuning the Nemeth and Kennedy [2] and Elsharkawy [3] correlations is shown in Equation (1). The summation of absolute error each time during training was set to be the objective function to be minimized. The Nelder–Mead algorithm [24,25], a commonly applied numerical method for minimization in multidimensional space, was applied to minimize the objective function, as shown in Equation (3).
Obj ( a ) = i = 1 n train | p dew , correlation ( Z l , T i , MW C 7 + , i , γ C 7 + | a ) p dew , i |
where a is the vector of the correlation coefficient for dew point pressure.
The outcomes of the correlations by Nemeth and Kennedy [2], Elsharkawy [3], Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5] for dew point pressure prediction are shown in Equations (12)–(15). After using Equations (12) and (15), we used the Nelder–Mead algorithm [24] to minimize the objective function in Equation (11) to tune the 11 coefficients of Nemeth and Kennedy [2], 19 coefficients of Elsharkawy [3], 28 coefficients of Ahmadi and Elsharkawy [4], and 19 coefficients of El-hoshoudy et al. [5], thus updating these correlations.
Nemeth and Kennedy [2] correlation:
l n p d = a 1 × ( z C 2 + z C O 2 + z H 2 S + z C 6 + 2 × ( z C 3 + z C 4 ) + z C 5 + 0.4 × z C 1 + 0.2 × z N 2 ) + a 2 × γ C 7 + + a 3 × ( z C 1 / ( z C 7 + + 0.002 ) ) + a 4 × T R + a 5 × ( z C 7 + × M W C 7 +   ) + a 6 × ( z C 7 + × M W C 7 +   ) 2 + a 7 × ( z C 7 + × M W C 7 +   ) 3 + a 8 × ( M W C 7 +   / ( γ C 7 + + 0.0001 ) ) + a 9 × ( M W C 7 +   / ( γ C 7 + + 0.0001 ) ) 2 + a 10 × ( M W C 7 +   / ( γ C 7 + + 0.0001 ) ) 3 + a 11
Elsharkawy [3] correlation:
p d = a 0 + a 1 × T f + a 2 × z H 2 S + a 3 × z C O 2 + a 4 × z N 2 + a 5 × z C 1 + a 6 × z C 2 + a 7 × z C 3 + a 8 × z C 4 + a 9 × z C 5 + a 10 × z C 6 + a 11 × z C 7 + + a 12 × M W C 7 +   + a 13 × γ C 7 + + a 14 × ( z C 7 + × M W C 7 + ) + a 15 × ( M W C 7 +   / γ C 7 + ) + a 16 × ( z C 7 + × M W C 7 +   / γ C 7 + ) + a 17 × ( z C 7 + / ( z C 1 + z C 2 ) ) + a 18 × ( z C 7 + / ( z C 3 + z C 4 + z C 5 + z C 6 ) )
Ahmadi and Elsharkawy [4] correlation:
p d = a 1 T f × z C 1 × a 2 + T f 2 × a 3 + z C 1 × a 4 + A × a 5
where
A = a 6 z N 2 1 3 × B × a 7 + z N 2 1 3 × C × a 8 + z N 2 2 3 × a 9 + B × a 10 + B 2 × a 11 + C × a 12
B = a 13 + M W C 7 +   3 × a 14 M W C 7 +   3 × z C 7 + 1 3 × a 15 M W C 7 +   6 × a 16 z C 4 1 3 × z C 7 + 1 3 × a 17 + z C 4 2 3 × a 18 + z C 7 + 1 3 × a 19 z C 7 + 2 3 × a 20
C = a 21 z C 1 × a 22 + z C 1 × γ C 7 + 3 × a 23 + z C 1 × z C 7 + 1 3 × a 24 + z C 1 2 × a 25 γ C 7 + 3 × a 26 + γ C 7 + 3 × z C 7 + 1 3 × a 27 z C 7 + 2 3 × a 28
El-hoshoudy et al. [5] correlation:
p d = x 3 + y 3
where,
x = a 0 + a 1 z C 1 + a 2 z C 2 + a 3 z C 4 + a 4 z C 4 2 + a 5 z C 5 + a 6 z C 6 + z C 7 + + a 7 T f a 8 T f + a 9
y = b 0 + b 1 z C 1 M W C 7 + + b 2 z C 3 + b 3 z C 7 + + b 4 ln z C 7 + + b 5 z C O 2 + b 6 z N 2 b 7 γ C 7 + + b 8
The cross-plots of the proposed model (XGBoost [12,13] with pseudocritical property correlations and artificial proxy features adopted from Sutton [22]) and four empirical correlations are shown in Figure 3. It can be observed that the proposed model performs better than the other models even after the recalibration of the existing correlation coefficients of the subject correlations, as shown in Figure 3a. As seen in Figure 3c, the overall performance of the tuned Ahmadi and Elsharkawy [4] correlation is better than the other three correlations. However, the tuned Ahmadi and Elsharkawy [4] correlation has fewer overprediction problems when predicting dew point pressure higher than 8000 psia. In addition, based on Figure 3d, the tuned Elsharkawy [3] correlation performs better than the tuned Nemeth and Kennedy [2] correlation and the tuned El-hoshoudy et al. [5] correlation. As shown in Figure 3e, the tuned El-hoshoudy et al. [5] correlation has better prediction compared to the Nemeth and Kennedy [2] correlation but still has a large range of prediction errors for the overall range of dew point pressure. As shown in Figure 3f, the tuned Nemeth and Kennedy [2] correlation has also underprediction problems for less than 3000 psia while having overprediction problems for pressures higher than 6000 psia.
The distribution of the errors for all five methods is shown in Figure 4. The proposed model is still the best method compared against the other correlations, as shown in Figure 4a. It has a higher density in prediction errors between −200 psia and 200 psia. The predicted errors of the developed model are in the range between −1400 psia and 1400 psia, and only a few samples are predicted errors outside of these ranges. The predicted errors of all four correlations are in the range between −3000 psia and 3000 psia. It is also obvious that the developed model has the highest peak density around zero. Others have similar peak densities around the predicted error of zero psia; the higher the peak density around the predicted error of zero psia, the better the performance of the developed model.
The error statistics of the five model performances are shown in Table 3, and based on the results, the proposed model has the best performance. The deviation of mean relative errors after adding pseudocritical property correlations of Sutton [22] and artificial proxy features into XGBoost [12,13] improves with respect to XGBoost [12,13] on the average of 35 psia. The proposed model performs better than others with a mean relative error of 470 psia (mean absolute relative error of 7.16%). The three correlations ([3,4,5]) for dew point pressure have similar error statistics with respect to each other, and their performance is better than the previously published correlations of Nemeth and Kennedy [2] (See Table 4). Their mean absolute error is in the range of 678 psia to 722 psia (mean absolute relative error of 10.69–11.36%), while the Nemeth and Kennedy [2] correlation has mean absolute errors of 838 psia (mean absolute relative error of 13.15%).

5. Summary and Conclusions

In this study, we reviewed a number of reports, and only higher-quality data was utilized (leading to about 350 points). Some of the fluids had excessive contamination, and some of them did not have reliable molecular weights and/or compositions, or overall quality check via modeling indicated issues with the data quality. Compositional information such as mole percentages of hydrocarbons to heptane plus and non-hydrocarbon gases, carbon dioxide, nitrogen, and hydrogen sulfide, along with the molecular weight of well stream, molecular weight, and specific gravity of heptane plus and reservoir temperature were used as predictors in order to estimate the dew point pressures. The samples, as reflected by their compositional character, are representative of gas condensates throughout the world with reservoir temperatures ranging from 60 °F to 381 °F and measured dew point pressures ranging from 1465 psia to 11,815 psia.
In developing the proposed model, an advanced machine learning algorithm, XGBoost [12,13,26,27], was performed on the system. Then, XGBoost [12,13,26,27] with Sutton [22] correlations for gas pseudocritical properties and eight artificial proxy features were selected based on the feature correlative responses on dew point pressure measurements. Higher-ranked features were applied to the datasets.
The outcomes of our study were compared against the tuned correlations of Nemeth and Kennedy [2], Elsharkawy [3], Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5]. In general, high-molecular-weight substances have very pronounced effects on the dew point pressure, as their behavior considering the dew point pressure is influenced by their quantity and as well as distribution. However, providing compositions to the heptane plus level reduces the predictive capabilities of the proposed model (and as well as the correlations), i.e., the compositions are not granular enough. However, such detailed compositions are not always available, especially in the older reports, and also sometimes they may not be very accurate. Therefore, in this study, we used compositional details to the heptane plus level as in the case of the classical correlations.
Some of the key conclusions of this study are as follows:
  • XGBoost [12,13,26,27] with gas pseudocritical property correlations of Sutton [22] and artificial proxy features leads to the best overall estimate of dew point pressure with the mean relative error of 470 psia (with mean absolute relative error = 7.16%) when compared to the tuned correlations of Nemeth and Kennedy [2], Elsharkawy [3], Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5].
  • After tuning, the correlations by Elsharkawy [3], Ahmadi and Elsharkawy [4], and El-hoshoudy et al. [5] for dew point pressure perform better than the Nemeth and Kennedy [2] correlation for dew point pressure. However, they all exhibit higher predictive errors for the elevated dew point pressure ranges (>8000 psia), while the proposed model performs better than all the selected correlations.
In conclusion, the implementation of the proposed model for precise dew point pressure prediction significantly bolsters phase-equilibrium modeling, fostering enhanced comprehension and control of fluid dynamics within reservoirs, pipelines, and processing facilities. This advancement notably refines decision-making processes pertaining to fluid management and process optimization. The meticulous determination of dew point pressure profoundly benefits reservoir engineering endeavors, facilitating comprehensive reservoir characterization and optimal production strategy formulation. By meticulously forecasting dew point pressure, operators are empowered to optimize natural gas extraction processes, thereby ensuring maximal hydrocarbon recovery with minimal energy expenditure and operational overheads and augmenting overall production efficacy. Furthermore, an adept grasp of dew point pressure is instrumental in emission mitigation efforts, as it enables the fine-tuning of gas processing and handling protocols, thereby curbing the release of greenhouse gases and pollutants into the atmosphere and advancing environmental sustainability. Moreover, precise dew point pressure estimation is indispensable for CO2 capture and storage initiatives, underpinning the advancement of sustainable carbon capture, utilization, and storage (CCUS) endeavors aimed at combatting climate change by mitigating greenhouse gas emissions.

Author Contributions

Conceptualization, T.L., B.D. and U.S.; methodology, T.L., B.D., L.L., X.Y. and U.S.; software, T.L., B.D., L.L. and X.Y.; validation, B.D. and X.Y.; formal analysis, L.L. and X.Y.; investigation, T.L. and L.L.; resources, B.D.; data curation, L.L.; writing—original draft, T.L. and X.Y.; writing—review & editing, U.S.; visualization, T.L.; supervision, B.D. and U.S. All authors have read and agreed to the published version of the manuscript.

Funding

The data that have been used are confidential.

Data Availability Statement

The data that have been used are confidential.

Acknowledgments

The authors express gratitude to Shell International Exploration and Production Inc., for authorizing the publication of this work. They also extend appreciation to Keat Hoe Foo from Shell for his assistance with the data. Utkarsh Sinha, a Volunteer Research Associate at the University of Houston acknowledges the members of the Research Consortium on Interaction of Phase Behavior and Flow (IPB&F) Consortium, along with its affiliates at the University of Houston.

Conflicts of Interest

Authors Xi Yang (new address: Citibank, 3800 Citibank Center, Tampa, FL 33610, USA) and Ligang Lu were employed by were employed by the company Shell International E&P Inc., Westhollow Technology Center. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Nomenclature

γ g Specific gravity of gas
γ C 7 + Specific gravity of heptane plus
a The vector of coefficients of correlation for dew point pressure
a i Constant, i = 1, 2, …, n
APFArtificial proxy feature
b i Constant, i = 1, 2, …, n
GEPGene expression programming
GORGas–oil ratio
LightGBMLight gradient boosting machine
MWMolecular weight, g/mole
M W C 7 + Molecular weight of heptane plus, g/mole
NNNeural network
p d p d e w = Dew point pressure, psia
PpcPseudocritical pressure, psia
P p c _ c o r r Pseudocritical pressure correction
SGSpecific gravity
S G C 7 + Specific gravity of heptane plus
S G C n + Specific gravity of carbon n plus
TRReservoir temperature, °R
TfReservoir temperature, °F
TpcPseudocritical temperature, °R
T p c _ c o r r Pseudocritical temperature correction
T r e s Reservoir temperature, °F
XGBoostExtreme gradient boosting
z C 1 Mole percent of methane
z C 2 Mole percent of ethane
z C 3 Mole percent of propane
z i C 4 Mole percent of iso-butane
z n C 4 Mole percent of normal butane
z i C 5 Mole percent of iso-pentane
z n C 5 Mole percent of normal pentane
z C 6 Mole percent of hexane
z C 7 + Mole percent of heptane plus
z C O 2 Mole percent of carbon dioxide
z H 2 S Mole percent of hydrogen sulfide
z i Mole percent of component i
z N 2 Mole percent of nitrogen

Appendix A

The cross-plots of four advanced machine learning models are shown in Figure A1. From Figure A1a,b, it can be inferred that XGBoost and LightGBM perform similarly, as expected, while XGBoost has slightly better accuracy in the high dew point pressure range. Random forest regressor shows obvious overprediction in the low range of dew point pressure (<4000 psia) and underprediction in the high range of dew point pressure (>6000 psia), as shown in Figure A1c. NN regressor with eight neurons in one hidden layer performs the worst among the four machine learning algorithms, as shown in Figure A1d. When the training dataset has a limited size, neural network algorithms more easily overfit the training data, and this is the significant problem causing the worse performance of the NN regressor. The error distribution plot shown in Figure A2 confirms the observation in Figure A1 that XGBoost is the best model among the four models. XGBoost has only a few samples with prediction errors outside the range from −2000 to 2000 psia. The prediction error statistical results are shown in Table A1. It is also obvious from Figure A2 that XGBoost performs the best, with a mean absolute error of 505 psia (mean absolute relative error of 8.196%), while among the other similar advanced machine learning algorithms, LightGBM becomes the second best, with a mean absolute error of 517 psia (mean absolute relative error of 8.361%). All machine learning models still cannot yield a closer estimation considering the predicted dew point pressure relative to the measured dew point pressure. Therefore, correlations of gas pseudocritical properties and artificial proxy features were added to the selected machine learning model, XGBoost, not only reducing the model error but also improving the model performance.
Figure A1. Cross-plots of measured p d e w vs. predicted p d e w using (a) XGBoost; (b) LightGBM; (c) random forest regressor; (d) NN regressor (8 neurons in 1 hidden layer).
Figure A1. Cross-plots of measured p d e w vs. predicted p d e w using (a) XGBoost; (b) LightGBM; (c) random forest regressor; (d) NN regressor (8 neurons in 1 hidden layer).
Fuels 05 00030 g0a1
Figure A2. Prediction error distribution plots of (a) XGBoost; (b) LightGBM; (c) random forest regressor; (d) NN regressor (8 neurons in 1 hidden layer).
Figure A2. Prediction error distribution plots of (a) XGBoost; (b) LightGBM; (c) random forest regressor; (d) NN regressor (8 neurons in 1 hidden layer).
Fuels 05 00030 g0a2
Table A1. Ten-fold cross-validation prediction error statistics.
Table A1. Ten-fold cross-validation prediction error statistics.
QuantityXGBoostLightGBMRandom ForestNeural Network
Mean absolute relative error (%)8.1968.3619.51313.212
Mean absolute error (psia)505517592705
Absolute relative error standard deviation (%)9.6009.15411.48637.524
Mean relative error (%)1.8742.0792.8124.864
Relative error standard deviation (%)12.48312.22214.64639.484
Pearson correlation0.9520.9510.9340.831
R-square0.9050.9030.8720.661

References

  1. Sage, B.H.; Olds, R.H. 1947. Volumetric Behavior of Oil and Gas from Several San Joaquin Valley Fields. Trans. AIME 2011, 170, 156–173. [Google Scholar] [CrossRef]
  2. Nemeth, L.K.; Kennedy, H.T. A Correlation of Dewpoint Pressure with Fluid Composition and Temperature. Soc. Pet. Eng. J. 1967, 7, 99–104. [Google Scholar] [CrossRef]
  3. Elsharkawy, A.M. Characterization of the Plus Fraction and Prediction of the Dewpoint Pressure for Gas Condensate Reservoirs. In Proceedings of the SPE Western Regional Meeting, Bakersfield, CA, USA, 26–30 March 2001. [Google Scholar]
  4. Ahmadi, M.A.; Elsharkawy, A. Robust correlation to predict dew point pressure of gas condensate reservoirs. Petroleum 2016, 3, 340–347. [Google Scholar] [CrossRef]
  5. El-hoshoudy, A.N.; Gomaa, S.; Desouku, S.M. Prediction of Dew Point Pressure in Gas Condensate Reservoirs Based on a Combination of Gene Expression Programming (GEP) and Multiple Regression Analysis. Pet. Petrochem. Eng. J. 2018, 2, 000163. [Google Scholar]
  6. Jalali, F.; Abdy, Y.; Akbari, M.K. Using Artificial Neural Network’s Capability for Estimation of Gas Condensate Reservoir’s Dew Point Pressure. In Proceedings of the EUROPEC/EAGE Conference and Exhibition, London, UK, 11–14 June 2007. [Google Scholar]
  7. Alzahabi, A.; El-Banbi, A.; Trindade, A.A.; Soliman, M. A regression model for estimation of dew point pressure from down-hole fluid analyzer data. J. Pet. Explor. Prod. Technol. 2017, 7, 1173–1183. [Google Scholar] [CrossRef]
  8. Alarfaj, M.K.; Abdulraheem, A.; Busaleh, Y.R. Estimating Dewpoint Pressure Using Artificial Intelligence. In Proceedings of the SPE Saudi Arabia Section Young Professionals Technical Symposium, Dhahran, Saudi Arabia, 19–21 March 2012. [Google Scholar]
  9. Brownlee, J. How Much Training Data Is Required for Machine Learning? Machine Learning Mastery. 2017. Available online: https://machinelearningmastery.com/much-training-data-required-machine-learning/ (accessed on 9 November 2020).
  10. Maheswari, J.P. Breaking the Curse of Small Data Sets in Machine Learning: Part 2 How Data Size Impacts Deep Learning Models and How to Work with Small Datasets? Towards Data Science. 2019. Available online: https://towardsdatascience.com/breaking-the-curse-of-small-data-sets-in-machine-learning-part-2-894aa45277f4 (accessed on 10 November 2023).
  11. XGBDocs. Available online: https://xgboost.readthedocs.io/en/stable/python/index.html (accessed on 2 April 2022).
  12. Chen, R.-C.; Caraka, R.E.; Arnita; Goldameir, N.E.; Pomalingo, S.; Rachman, A.; Toharudin, T.; Tai, S.-K.; Pardamean, B. An end to end of scalable tree boosting system. Sylwan 2020, 165, 1–11. [Google Scholar]
  13. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2016 (KDD ‘16), San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  14. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  15. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  16. Sinha, U.; Dindoruk, B.; Soliman, M.Y. Physics augmented correlations and machine learning methods to accurately calculate dead oil viscosity based on the available inputs. SPE J. 2022, 27, 3240–3253. [Google Scholar] [CrossRef]
  17. Sinha, U.; Dindoruk, B.; Soliman, M. Physics guided data-driven model to estimate minimum miscibility pressure (MMP) for hydrocarbon gases. Geoenergy Sci. Eng. 2023, 224, 211389. [Google Scholar] [CrossRef]
  18. Available online: https://www.geomarkresearch.com/rfdbase (accessed on 2 April 2022).
  19. Shokir, E.M. Dewpoint Pressure Model for Gas Condensate Reservoirs Based on Genetic Programming. Energy Fuels 2008, 22, 3194–3200. [Google Scholar] [CrossRef]
  20. Kaydani, H.; Mohebbi, A.; Hajizadeh, A. Dew point pressure model for gas condensate reservoirs based on multi-gene genetic programming approach. Appl. Soft Comput. 2016, 47, 168–178. [Google Scholar] [CrossRef]
  21. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd ed.; Springer: New York, NY, USA, 2009. [Google Scholar]
  22. Sutton, R.P. Fundamental PVT Calculations for Associated and Gas/Condensate Natural Gas Systems. In SPE Annual Technical Conference and Exhibition? SPE: New York, NY, USA, 2005; p. 97099. [Google Scholar]
  23. Matthews, T.A.; Roland, C.H.; Katz, D.L. High pressure gas measurement. In Proceedings of the Twenty-first Annual Convention, NGAA; 1942. [Google Scholar]
  24. Nelder, J.A.; Mead, R. A simple method for function minimization. Comput. J. 1965, 7, 308–313. [Google Scholar] [CrossRef]
  25. Gao, F.; Han, L. Implementing the Nelder-Mead simplex algorithm with adaptive parameters. Comput. Optim. Appl. 2012, 51, 259–388. [Google Scholar] [CrossRef]
  26. Swalin, A. CatBoost vs. Light GBM vs. XGBoost. Towards Data Science. 2018. Available online: https://towardsdatascience.com/catboost-vs-light-gbm-vs-xgboost-5f93620723db (accessed on 28 June 2024).
  27. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
Figure 1. Brief Summary of the loss function minimized in XGBboost [12,13].
Figure 1. Brief Summary of the loss function minimized in XGBboost [12,13].
Fuels 05 00030 g001
Figure 2. Feature correlation with dew point pressure heatmap for XGBoost [9] with Sutton [22] pseudocritical properties.
Figure 2. Feature correlation with dew point pressure heatmap for XGBoost [9] with Sutton [22] pseudocritical properties.
Fuels 05 00030 g002
Figure 3. Cross-plots of measured p d e w vs. predicted p d e w using (a) XGBoost [9] with Sutton [22] pseudocritical properties and APF; (b) XGBoost [12,13]; (c) tuned Ahmadi and Elsharkawy [4] correlation; (d) tuned Elsharkawy [3] correlation; (e) tuned El-hoshoudy et al. [5] correlation; (f) tuned Nemeth and Kennedy [2] correlation.
Figure 3. Cross-plots of measured p d e w vs. predicted p d e w using (a) XGBoost [9] with Sutton [22] pseudocritical properties and APF; (b) XGBoost [12,13]; (c) tuned Ahmadi and Elsharkawy [4] correlation; (d) tuned Elsharkawy [3] correlation; (e) tuned El-hoshoudy et al. [5] correlation; (f) tuned Nemeth and Kennedy [2] correlation.
Fuels 05 00030 g003
Figure 4. Prediction error distribution plots using (a) XGBoost [12,13] with Sutton [22] pseudocritical properties and APF; (b) XGBoost [12,13]; (c) tuned Ahmadi and Elsharkawy [4] correlation; (d) tuned Elsharkawy [3] correlation; (e) tuned El-hoshoudy et al. [5] correlation; (f) tuned Nemeth and Kennedy [2] correlation.
Figure 4. Prediction error distribution plots using (a) XGBoost [12,13] with Sutton [22] pseudocritical properties and APF; (b) XGBoost [12,13]; (c) tuned Ahmadi and Elsharkawy [4] correlation; (d) tuned Elsharkawy [3] correlation; (e) tuned El-hoshoudy et al. [5] correlation; (f) tuned Nemeth and Kennedy [2] correlation.
Fuels 05 00030 g004
Table 1. Dataset (342 samples) statistical summaries.
Table 1. Dataset (342 samples) statistical summaries.
MeanStandard DeviationMinMedianMax
Reservoir Temperature (°F)219.0450.7460.00217.00381.00
Dew Point Pressure (psia)685522231465664811815
N 2 (mole %)0.500.800.000.3010.15
C O 2   (mole %)2.283.520.000.4733.72
H 2 S (mole %)0.070.970.000.0016.83
C 1 (mole %)77.7510.2544.6276.5399.06
C 2 (mole %)5.782.730.195.6414.15
C 3 (mole %)3.151.550.053.1112.36
i C 4 (mole %)0.640.330.010.494.36
n C 4 (mole %)1.260.630.001.283.73
i C 5 (mole %)0.520.330.010.494.36
n C 5 (mole %)0.600.310.000.601.85
C 6 (mole %)0.840.520.010.844.81
C 7 + (mole %)6.613.880.096.6917.08
Molecular Weight of C 7 + (g/mole)174.6834.08106.00170.81283.16
Specific Gravity of C 7 + 0.810.030.720.810.90
Reservoir Fluid MW (g/mole)31.959.4716.4031.1365.97
Table 2. Feature correlation distribution with dew point pressure using XGBoost [12,13] with pseudocritical properties according to Sutton [22].
Table 2. Feature correlation distribution with dew point pressure using XGBoost [12,13] with pseudocritical properties according to Sutton [22].
FeaturesCorrelation with Pdew
Dewpoint Pressure1.0000
Molecular Weight of C 7 + 0.7539
Specific Gravity of C 7 + 0.7041
Mole % of C 2 0.3741
Mole % of C 7 + 0.3186
T p c 0.3140
P p c 0.3059
Reservoir Fluid MW0.3040
Mole % of C 3 0.2854
Mole % of C O 2 0.2839
Mole % of N 2 0.2413
Mole % of i C 4 0.1912
Mole % of n C 4 0.1665
Reservoir Temperature0.1473
Mole % of H 2 S 0.1261
Mole % of i C 5 0.1139
Mole % of n C 5 0.0340
Mole % of C 6 0.0015
Table 3. Configuration of XGBoost [12,13] method used in this work.
Table 3. Configuration of XGBoost [12,13] method used in this work.
InputsReservoir Temperature (°F);
mole % of N 2 , C O 2 , C 1 , C 2 , C 3 , n C 4 , i C 4 , n C 5 , i C 5 , C 6   C 7 + ;
M W C 7 + ;   S G C 7 + ;
T p c ; P p c ;
APF#1, 2, 3, 4, 5, 6, 7, 8
Test–Train Split30–70%
Cross Validation Folds (Training Data)10
BoosterGbtree
Objectivereg:squarederror
learning_rate0.01
max_depth6
min_child_weight1
Subsample0.5
colsample_bytree0.5
reg_alpha23
Reg_lambda0.5
Gamma75
+ The description of all configuration hyperparameters can be found in XGBDocs, 2022 [11].
Table 4. Ten-fold cross-validation prediction error statistical results.
Table 4. Ten-fold cross-validation prediction error statistical results.
QuantityXGBoost [12,13] with Sutton [22] Pseudocritical Properties and APFXGBoost [12,13]Tuned Ahmadi and Elsharkawy [4]Tuned Elsharkawy [3]Tuned El-hoshoudy et al. [5]Tuned Nemeth and Kennedy [2]
Mean absolute relative error (%)7.1608.19610.69311.07511.35913.149
Mean absolute error (psia)470505678722715838
Absolute relative error standard deviation (%)7.2289.60010.74610.05811.46112.625
Mean relative error (%)1.2221.8740.007−0.2030.9844.650
Relative error standard deviation (%)10.10112.48315.16014.96016.10717.625
Pearson correlation0.9570.9520.9120.9020.9040.880
R-square0.9160.9050.8240.8090.8120.736
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lertliangchai, T.; Dindoruk, B.; Lu, L.; Yang, X.; Sinha, U. A Comparative Analysis of the Prediction of Gas Condensate Dew Point Pressure Using Advanced Machine Learning Algorithms. Fuels 2024, 5, 548-563. https://doi.org/10.3390/fuels5030030

AMA Style

Lertliangchai T, Dindoruk B, Lu L, Yang X, Sinha U. A Comparative Analysis of the Prediction of Gas Condensate Dew Point Pressure Using Advanced Machine Learning Algorithms. Fuels. 2024; 5(3):548-563. https://doi.org/10.3390/fuels5030030

Chicago/Turabian Style

Lertliangchai, Thitaree, Birol Dindoruk, Ligang Lu, Xi Yang, and Utkarsh Sinha. 2024. "A Comparative Analysis of the Prediction of Gas Condensate Dew Point Pressure Using Advanced Machine Learning Algorithms" Fuels 5, no. 3: 548-563. https://doi.org/10.3390/fuels5030030

Article Metrics

Back to TopTop