Next Article in Journal
A State Transition Diagram and an Artificial Physarum polycephalum Colony Algorithm for the Flexible Job Shop Scheduling Problem with Transportation Constraints
Previous Article in Journal
Optimization of Batch Crystallization of Magnetic Lysozyme Crystals and Study of the Continuous Crystallization Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CLAP: Gas Saturation Prediction in Shale Gas Reservoir Using a Cascaded Convolutional Neural Network–Long Short-Term Memory Model with Attention Mechanism

1
Shale Gas Institute of PetroChina Southwest Oil & Gasfield Company, Chengdu 610051, China
2
Research Institute of Petroleum Exploration and Development, PetroChina, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Submission received: 9 August 2023 / Revised: 31 August 2023 / Accepted: 1 September 2023 / Published: 4 September 2023

Abstract

:
Gas saturation prediction is a crucial area of research regarding shale gas reservoirs, as it plays a vital role in optimizing development strategies and improving the efficiency of exploration efforts. Despite the advancements in deep learning techniques, accurately modeling the complex nonlinear relationships involved in gas saturation prediction remains a challenge. To address this issue, we propose a novel cascaded model, CLAP, combining convolutional neural networks (CNNs) and Long Short-Term Memory (LSTM) with an attention mechanism. It effectively captures and visualizes the intricate nonlinear relationships, enabling accurate gas saturation prediction in shale gas reservoirs. In this study, nine logging curves from 27 shale gas wells in the Changning area of the Sichuan Basin were used to train the CLAP model for predicting the gas saturation of the Wufeng-Longmaxi Formation shale. Compared to the Archie and random forest models, the CLAP model exhibited enhanced accuracy in predicting shale gas saturation. Promisingly, the CLAP model demonstrates outstanding statistical performance in gas saturation prediction, achieving an impressive R2 score of 0.762 and a mean square error (MSE) score of 0.934. These positive results highlight the effectiveness and potential utility of our proposed CLAP model in accurately predicting gas saturation in shale gas reservoirs. The application of deep learning techniques, such as CNNs, LSTM, and attention mechanisms, presents a promising avenue for further advancements in this field.

1. Introduction

A shale gas reservoir is a very special type of reservoir, and the prediction of its gas saturation is very important for evaluating the gas reserves and development potential of the reservoir. However, due to the complexity and particularity of shale gas reservoirs, the prediction of gas saturation faces some challenges. Firstly, the porosity of shale gas reservoirs is very low, and the gas mainly exists in the micropores and nanopores of the rock in the adsorbed state. The release and flow mechanism of this adsorbed gas is very different from that of traditional reservoirs, so the traditional gas saturation model may not be suitable for shale gas reservoirs. Secondly, the rock properties and components of shale gas reservoirs are complex and diverse, including organic matter content, general rock texture (e.g., interlocking pattern, shales, and habits of primary mineral grains), mineral composition, and pore structure. The influence of these factors on gas saturation is very complicated, and traditional models often fail to accurately consider the influence of these factors. Researchers have proposed new methods and models to predict shale gas reservoirs’ gas saturation. These methods include the electrical resistivity method, the non-electro special technique, and deep learning methods. These methods attempt to comprehensively consider factors such as rock properties, geological conditions, and stratigraphic characteristics to improve the prediction accuracy of gas saturation.
The electrical resistivity method was born in the 1920s, and was mainly used for the identification of oil and gas reservoirs, lithology classification, and quantitative interpretation and evaluation of reservoirs [1,2,3,4,5]. However, the study of water saturation did not make any progress until Archie proposed the calculation method of water saturation in 1942 [2]. While the Archie formula has been utilized, it comes with stringent prerequisites and challenging conditions regarding formation [3,4]. A succession of resistivity–porosity–saturation models for water saturation measurement have been developed, with several earning recognition in shale applications. Literature reviews indicate that primary factors affecting moisture estimation include negatively charged clays, clay swelling, conductive kerogen, pyrite, adsorbed water, pore structure, and low permeability [1,2,3,4,5,6,7]. Many studies have been carried out on the relationship between resistivity logging data and water saturation in sandstone reservoirs, especially focusing on the Simandoux model [3], the improved Simandoux model [4], and the Indonesian model [5]. Researchers have improved these models to analyze the relationship between water saturation and different types of clay minerals in more depth, such as the dual-water model [6], the Clay model, and the Waxman–Smits–Thomas (WST) model [7,8,9]. More recently, researchers have also conducted many studies on water saturation prediction in organic-rich shales [10,11,12]. The advantages of the electrical resistivity method are that it is simple to use, highly reliable, and suitable for saturation prediction under general geological conditions. However, shale reservoirs differ from conventional sandstone reservoirs. Factors such as the graphitization of organic matter, high pyrite content, and the added conductivity from clay minerals have a pronounced impact on resistivity. In the resistivity method (electrical approach), the more variable the parameters included in the response equation, the greater the iterative solution error. Traditional electrical logging saturation calculation methods are not particularly suitable for low-resistivity shale gas reservoirs.
In recent years, numerous scholars have conducted research on non-electrical saturation evaluations of shale gas reservoirs using both conventional and specialized logging data [5,6,7,8,9,10]. At present, some researchers relate water saturation to shale skeleton minerals, such as quartz, clay minerals, feldspar, and carbonates (including calcite, dolomite, and siderite) through statistical regression to obtain empirical formulas [10,12,13,14,15,16]. Their models may only be applicable to specific areas, but they work well [14,15,16]. The electrical method of predicting water saturation is common in other countries, whereas the non-electrical method is common in China’s more complex and unique geological conditions [12,13,14,15,16]. Zhang proposed a model that considers different mineral compositions through statistical regression [12]. Yan, based on the fact that the water saturation of shale reservoirs in the Sichuan Basin decreases with the decrease in clay mineral content and the increase in organic matter, indicates that the calculation model with dual factors is more appropriate [13]. Shi proposed a method based on the ratio of the background value of organic matter to the measured and calculated value of TOC, but did not consider the deviation effect of bound water and TOC in clay on the calculation accuracy of water saturation [14]. Li found that the bound water saturation of shale reservoirs is closely related to the clay mineral content [15,16]. To sum up, there is an urgent need to develop a gas saturation prediction model that is widely suitable for most shale reservoirs. Yet, traditional methodologies exhibit distinct limitations [8,9,10,11,12,13,14]. The non-electrical logging method faces challenges such as inaccuracies in the indirect calculation of key reservoir parameters from logging data and a limited amount of experimental data available for correcting these computed parameters [12,13,14,15,16]. Additionally, for shales with diverse mineral combinations, mineral structures, and reservoir properties, it becomes challenging to apply a single non-electrical model to predict gas saturation across different types of low-resistivity shales. It is imperative to develop new gas saturation prediction models based on the mineral composition, organic maturity, and unique logging data characteristics of each specific type of low-resistivity shale [12,13,14,15,16].
In the field of petroleum science and engineering, it has become a common method to use machine learning and deep learning algorithms for logging data analysis and prediction. Aifa successfully predicted the permeability and porosity of the Hassi R’Mel gas field in Algeria using a neuro-fuzzy system [17]. Al-Mudhafar integrated advanced machine learning algorithms with log interpretation to accurately model lithofacies classification and permeability by deriving complex relationships in logging data [18]. Wood proposed a TOB learning network algorithm as an optimized data-matching algorithm with high accuracy and interpretability, which provides an ideal tool for deeper data mining [19]. Otchere introduced a novel ensemble model, combining random forest and Lasso regularization techniques for feature selection, which enhanced reservoir representation using the Extreme Gradient Boosting (XGBoost) regression model for permeability and water saturation prediction [20]. Drawing from the classical capillary pressure formula, Xu established a linear regression model linking rock’s capillary force inversely to porosity squared and correlated it with trap height in the pure gas zone [21]. Huang introduced a shale gas reservoir saturation evaluation model using the random forest regression algorithm, offering a highly adaptable tool with strong generalization for shale gas development. Huang introduced a shale gas reservoir saturation evaluation model using the random forest regression algorithm, offering a highly adaptable tool with strong generalization for shale gas development [22]. However, most machine learning models heavily rely on data preprocessing methods and often struggle to handle complex nonlinear relationships [23,24,25,26,27,28,29,30,31]. This limitation holds true for gas saturation prediction in shale gas reservoirs, where accurately capturing the intricate nonlinear connections across various reservoir indicators becomes crucial [32,33,34,35,36,37].
Historically, gas saturation estimation of the Wufeng-Longmaxi shale in the Southern Sichuan basin was primarily based on the resistivity model [38,39,40,41,42]. However, this method’s limited precision compromised the accuracy of reservoir evaluations [38,39,40,41,42]. In response, this study integrates well logs with experimentally measured gas saturation data, utilizing a cascaded deep learning framework that combines convolutional neural networks (CNNs), Long Short-Term Memory (LSTM), and attention mechanisms (ATTs) to predict the gas saturation of a shale gas reservoir. This approach seeks to harness the strengths of deep learning methodologies in tandem with geological rock and gas reservoir characteristics and logging data, aiming to bolster the precision and reliability of gas saturation predictions.

2. Saturation Calculation Methods

2.1. Processing of Data Set

2.1.1. Source of Data

In this study, we utilized a dataset consisting of nine logging curves, namely natural gamma (GR), uranium-free gamma (KTH), uranium (U), potassium (K), acoustic time difference (AC), neutron (CNL), density (DEN), deep lateral resistivity (RT), and shallow lateral resistivity (RXO), collected from Schlumberger. These curves correspond to the Wufeng-Longmaxi Formation shale in 27 shale gas wells located in the Changning area, located at the southern edge of the Sichuan Basin. To train our model, we employed a total of 1436 gas saturation measurements from these wells as the experimental data. Subsequently, we utilized this dataset to train the model, allowing it to learn the complex relationships between the logging curves and gas saturation levels. Following the training phase, we applied the trained model to predict the gas saturation levels in an additional set of five shale gas wells that were not included in the training phase.
All these samples were analyzed for water saturation. Based on these measurements, gas saturation was determined by subtracting the water saturation from 100%. Dry shale was put in an environment with constant humidity. Vapor diffusion absorption occurs in shale when the humidity in the shale is lower than that in the external environment, increasing the water content in the shale. Then, water saturation was calculated based on the quality variation of the shale samples during this period [25,26].

2.1.2. Analysis of Feature Correlation and Importance

Correlation analysis is a statistical method used to measure and evaluate the strength and direction of relationships between variables to reveal intrinsic connections, predict modeling, explore data, assist feature selection, and find patterns, trends, and anomalies to improve model efficiency and accuracy.
Feature correlations are visualized by plotting, where the numerical magnitude represents the degree of correlation, the maximum value is 1, and a plus or minus sign represents a positive or negative correlation. Since the main task of this experiment is to predict water saturation, we mainly look at the correlation between water saturation and other columns, so that we can select the features with high correlation for prediction. In the second column of the figure, we can see that the correlation of all columns is greater than 0.1, which means that the prediction of water saturation is relevant.
In order to reveal the degree of association between each feature and the target variable and to understand which features play a key role in the prediction of the target variable, the machine learning random forest model has an important feature that can analyze the importance of features [23]. Therefore, a random forest model is constructed to predict the importance of features. In the prediction process, the importance of the input parameters is analyzed, that is, the contribution value of the logging curve type to the prediction results of the random forest regression algorithm. Ten logging curves of natural gamma (GR), uranium-free gamma (KTH), uranium (U), thorium (TH), potassium (K), acoustic time difference (AC), neutron (CNL), density (DEN), deep lateral resistivity (RT), and shallow lateral resistivity (RXO) were selected as the input curves of the sample, and core water saturation (SW) was used as the output label (Figure 1). The relationship between the well logging curve and the core water saturation was analyzed. It was found that DEN, CNL, RT, RXO, K, AC, GR, U, and KTH are of high importance in the prediction of water saturation in ultra-low resistivity shale gas reservoirs, while TH is of general importance. Therefore, nine well-logging curves with high importance were finally selected as the input curves of the prediction model.

2.2. Water Saturation Prediction Method

  • A brief description of the experimental procedure.
The process of this experiment is as follows: the original data set is obtained, and the experimental procedure involved several steps. Firstly, the original dataset was obtained and subjected to various analyses, including feature correlation and importance analysis, in order to explore the relationships within the data. The experimental model employed in this study is the CLAP model (combining convolutional neural networks (CNNs) and Long Short-Term Memory (LSTM) with an attention mechanism). The model underwent training using the prepared dataset, and its performance was evaluated. To assess the model’s effectiveness, it was utilized for gas saturation prediction. Figure 2 visually represents the complete pipeline for predicting water saturation using the CLAP model. It provides a clear illustration of the step-by-step process and elucidates the specific role played by each component within the pipeline. The CLAP model is a pluggable module bridging the split data to the gas saturation.
Furthermore, Table 1, Table 2 and Table 3 offer comprehensive comparisons and detailed analyses of the gas saturation prediction results, as well as the metrics among the three aforementioned models.

2.2.1. CNN Model

CNNs are particularly advantageous when dealing with data that exhibits a similar network structure, such as time series and image data. Their benefits include parameter sharing and sparse connections, which lead to a reduction in the number of learning parameters required. Consequently, CNNs can effectively train on smaller datasets, thereby mitigating the risk of overfitting. Usually, the CNN architecture is primarily comprised of several modules, with the convolutional layer being the central component that sets it apart from other neural networks. In this layer, filter parameters are employed to perform convolution operations on the input layer (experimental data), enabling the extraction of fundamental features. These filter parameters are initially randomly initialized, and then the defined loss function is utilized for backpropagation in order to obtain the most suitable filter parameters for feature extraction [24]. Taking the convolution of two-dimensional data as an example, the convolution operation formula is:
S i , j = X K i , j = m n X m , n K i m , j n
where * is the convolution operation.
The second module in a convolutional neural network is the pooling layer, which calculates aggregate statistics for a specific layer of the network. Its primary purpose is to reduce model size, improve computation speed, and enhance the robustness of extracted features to mitigate overfitting. The pooling layer has a significant advantage over the convolutional layer—it does not have any parameters to learn, thereby alleviating parameter pressure within the network. Two commonly used types of pooling layers are max pooling and average pooling (Table 1, Table 2 and Table 3). Let’s consider two-dimensional data as an example. Max pooling compresses the input data by selecting the maximum element value from the corresponding area, whereas average pooling calculates the average value within the area.
The third module is the fully connected layer, which performs a complete connection between neurons. It applies an appropriate activation function to generate the output activation values, representing the features extracted through the convolutional neural network [25].

2.2.2. LSTM Model

Long Short-Term Memory (LSTM) models are a variant of recurrent neural networks (RNNs), which are commonly used for modeling and prediction tasks that deal with sequence data (Figure 3). Compared with the traditional RNN model, the LSTM model introduces a gating mechanism, which can better capture and remember the long-term dependencies in the input sequence [26].
The core idea of LSTM model is to use memory units (cells) to store and transfer information, and to control the flow of information through gates. Specifically, the LSTM model consists of an input gate, a forget gate, an output gate, and a memory cell [27].
The input gate determines how much input information will be passed to the memory cell, the forget gate determines whether to delete the previous memory, and the output gate determines the output of the hidden state. The memory unit is responsible for storing and transferring long-term information, which is controlled and updated by the calculations of the gating unit.
Where: xt is the input data at time t; Ct−1 is the memory value at time t − 1; ht−1 is the output value of LSTM at time t − 1. The three data xt, Ct−1, and ht−1 constitute the input data of the model. Ct is the memory value at time t, ht is the output value of LSTM at time t, and the two data Ct and ht constitute the output data of the model [28].
Control functions of forget gate, input gate and output gate:
f i t = s i g m a b i f + j U i , j f x j t + j W i , j f h j t 1
g i t = s i g m a b i g + j U i , j g x j t + j W i , j g h j t 1
q i t = s i g m a b i o + j U i , j o x j t + j W i , j o h j t 1
where bo, Uc, Wo are the bias, the input weight, and the cyclic weight of the forget gate, respectively. In these variants, one can optionally use the cell state as an additional input to the three gates of the ith cell (Table 1, Table 2 and Table 3).

2.2.3. Attention Mechanism

Attention mechanism is an important component used in neural network models (Figure 4). Its principle is to assign different attention weights to different parts of the input data, thereby reducing the role of irrelevant parts [29,30]. This enables the model to focus more on important information during processing and learning tasks, ultimately enhancing performance. As shown in Figure 4, the attention mechanism focuses on relevant parts of the input sequence while generating an output. It involves computing attention weights that determine the importance of each element in the input sequence with respect to a query. These attention weights are used to calculate a context vector, which captures the most relevant information from the input. The context vector is then combined with the query to produce the final output.
So as to improve the performance and performance of the model, the relevant formula [29,30] is as follows:
α i = e s h i , h t j = 1 N e s h i , h t
α = i = 1 N α i h i
where α i is the score of the feature vector, and a higher score indicates greater attention. s h i , h t is the weight value of the ith input feature in the attention mechanism, which is the ratio of the score of the feature vector to the total population. Then, all vectors are summed and averaged to obtain the final vector α (Figure 4).

2.2.4. CLAP Model

The CNN-LSTM model with an attention mechanism is applied in the prediction of shale gas saturation. This study focuses on utilizing data from 27 shale gas wells located in the Changning area, situated on the southern margin of the Sichuan Basin, specifically targeting the ZhongWufeng Formation and Longmaxi Formation shale. The initial step of the model involves performing convolution operations on the input data, effectively extracting local features (depicted in Figure 5). The convolutional neural network (CNN) plays a crucial role in this extraction process. The extracted features obtained from the CNN are then fed into the Long Short-Term Memory (LSTM) network for sequential encoding. The LSTM network enables the modeling of long-term dependencies within the sequence data, generating a contextual representation. Following the LSTM output, attention weights are calculated to discern and emphasize significant segments of the input sequence. These attention weights serve as indicators of the relative importance of different parts of the sequence. By employing these weights, a weighted summation of the LSTM outputs is performed, obtaining an integrated context representation that better captures essential information. This integrated context representation subsequently serves as input to various structures, including the fully connected layer, facilitating the prediction of gas saturation.
The CLAP model offers several advantages. Firstly, the CNN component excels at extracting local features from the gas saturation experimental data, enabling effective capture of feature information at various scales (as demonstrated in Figure 6). This capability allows the model to extract multiple levels of abstract representation from the input sequence, thereby expressing the intrinsic characteristics of the experimental data more accurately. Secondly, the attention mechanism plays a pivotal role in the CLAP model. It selectively focuses on important segments of the input experimental data. Through self-learning, the model can autonomously determine which parts are crucial for predicting gas saturation, effectively introducing attention into the decision-making process. Lastly, due to the cohesive integration of the CNN, LSTM, and attention mechanism, the CLAP model exhibits strong expressive power and generalization ability [31]. It can effectively capture the information embedded within the input experimental data, leading to more precise predictions of gas saturation (as illustrated in Table 4 and Table 5).

2.3. Performance Evaluation

2.3.1. Evaluation Index of the Experiment

The mean absolute error (MAE) is also a common regression loss function, which indicates the average magnitude of error in the predicted value regardless of the direction of the error, and its formula is as follows:
M A E = 1 N i = 1 N | y j y ^ j |
Mean square error (MSE) is used in regression prediction tasks to measure the average squared difference between the predicted values of a model and the true values of the model. The formula is as follows:
M S E = 1 N i = 1 N y i y ^ j 2
The R2 score, also known as the coefficient of determination, is one of the statistical measures used to assess the goodness of fit of a regression model. It indicates the proportion of variance that the model is able to explain in the target variable, and its formula is as follows:
R 2 = 1 i = 1 n y i y ^ j 2 i = 1 n y i y j 2

2.3.2. Presentation and Analysis of Experimental Results

Based on the Pytorch framework, this experiment used the Python language to write code, and completed the construction and training of the CLAP combined model. The dataset was processed and divided into two parts: 70% of the data were used for training the experimental model, while the remaining 30% were utilized for model validation, and the data include feature depth, AC, GR, CNL, DEN.

2.3.3. Characteristics of Error Function Variation and Model Convergence

The CLAP model was trained using 1436 samples over 500 epochs. The training error variation throughout the training process is illustrated in Figure 7. An epoch, in this context, refers to the complete forward and backward propagation of all data through the network [13,14,15]. The loss plots indicated convergence after approximately 100 epochs. As the training progressed, the training error decreased to 0.014 (Figure 7a), while the R2 value steadily increased to 0.843 (Figure 7b). This suggests that the model’s regression analysis error was minimal, indicating good performance.

3. Result and Discussion

In this study, nine logging curves—including natural gamma (GR), uranium-free gamma (KTH), uranium (U), potassium (K), acoustic time difference (AC), neutron (CNL), density (DEN), deep lateral resistivity (RT), and shallow lateral resistivity (RXO)—were sourced from 27 shale gas wells in the Changning area at the southern edge of the Sichuan Basin. These curves, related to the Wufeng-Longmaxi Formation shale, were utilized to train the CLAP model. This trained model was then applied to predict gas saturation in five other shale gas wells not used in the training phase. Upon comparison with the Archie model (methods and details referenced in [13]) and the random forest model, the CLAP model demonstrated superior accuracy in predicting shale gas saturation.
Firstly, the prediction accuracy of the three prediction models was compared through the prediction results of gas saturation in a single well (Table 6, Figure 8). Well A is a shale gas well in the west of the Changning area. The lithology of the Wufeng-Longmaxi Formation is mainly black silica shale, gray calcarine shale, and gray clay shale, and the shale gas is well displayed. The GR curve of the Wufeng-Longmaxi Formation shows a high value at 1322–1313 m, and the lateral resistivity curve of the deep and shallow parts is low, ranging from 14 Ω·m to 77 Ω·m. The resistivity of some wells is less than 20 Ω·m, which makes the curve as a whole zigzag. By comparing the gas saturation prediction results of the Archie model, random forest model, and CLAP model in well A, it can be seen that the gas saturation prediction results of the CLAP model are obviously the most consistent with the experimental results of gas saturation obtained from the core. The gas saturation calculated by the Archie model is significantly lower than the experimental data of gas saturation. For the random forest model, the difference between the predicted and experimental results of gas saturation is significantly larger than that of the CLAP model. In order to further compare the accuracy of the prediction results of CLAP model, Archie model, and random forest model, the correlation analysis between the predicted gas saturation results of the three models and the core analysis of well A was carried out (Figure 9). It can be seen that the predicted results of the CLAP model have the strongest correlation with the experimental results, with an R2 of 0.97, while the predicted results of random forest model have the worst correlation with the experimental results, with an R2 of only 0.57.
The comparison of gas saturation prediction results of the three models with experimental data in the five shale gas wells is shown in Table 7. The mean absolute error (MAE), root mean square error (RMSE), and coefficient of determination (R2) between the predicted results of the CLAP model and the experimental data were 1.44–2.01, 0.89–1.87, and 0.74–0.98, respectively. It can be seen that there is a strong correlation between the predicted results and the experimental data. Compared with the Archie model, the MAE, RMSE, and R2 of the prediction results of the Archie model and the experimental data were between 12.34–27.43, 12.34–16.93, and 0.46–0.78, respectively, indicating a poor correlation between them. The MAE, RMSE, and R2 of the random forest model were between 8.45–16.45, 13.46–27.47, and 0.26–0.67, respectively, indicating the worst correlation. From the perspective of the three error evaluation parameters, the CLAP model has the smallest error between the prediction results and the experimental data. By comparing the calculation results of this well, it can be found that the calculation results of the Archie model are generally lower than the experimental results, and there are large differences in some data, while the prediction results of the random forest model are very different from the experimental results. Therefore, both the Archie model and the random forest model are not suitable for the calculation of shale gas saturation in the study area. However, the error between the prediction results of the CLAP model and the experimental results is relatively small, so the CLAP model is more suitable for the prediction of shale gas saturation in the Changning area (Figure 9).
Comparing the above three models, it can be seen that the CLAP model is superior to the traditional Archie model and the random forest model in the prediction calculation of shale gas saturation (Figure 9). Although the CLAP model does not have an equation representing physical properties, it can establish a more accurate prediction model of water saturation through multivariate data network training of logging parameters related to water saturation, and then obtain a more accurate gas saturation. The advantage of the CLAP model is that it can make full use of the diversity of logging data and combine the characteristics of convolutional neural networks (CNNs) and Long Short-Term Memory (LSTM) networks, has good memory ability in time series, and uses an attention mechanism (ATT) to improve the importance and correlation of features [43,44,45]. This enables the model to capture the time series information in the input data and automatically select the most relevant features for prediction [43,44,45]. The superior precision of the CLAP prediction model stems from two fundamental components: the incorporation of the attention mechanism and the synergistic combination of CNN and LSTM neural network architectures [46,47,48,49]. The attention mechanism plays a pivotal role by assigning varying attention weights to different segments of the input data. This enables the model to effectively diminish the influence of irrelevant information and instead prioritize relevant and important details during the processing and learning tasks [46,47,48,49]. The attention mechanism was integrated into CLAP, allowing it to selectively concentrate on significant aspects within the input experimental data [46,47,48,49]. By leveraging the attention mechanism, the model becomes adept at identifying and focusing on the most informative features in the input, leading to enhanced predictive capabilities [50,51]. Moreover, CNN and LSTM neural network structures are combined to model the implicit information of input data. The CNN is mainly responsible for extracting the local features of the experimental data of gas saturation, enabling it to effectively capture the feature information at different scales [50,51,52,53]. The LSTM component addresses the challenge of modeling long-term dependencies, which are crucial in sequence data [50,51,52,53]. Combining these two neural network structures results in a synergistic approach, enabling a more comprehensive understanding of the input experimental data, resulting in an improved prediction model [50,51,52,53].
Compared to the CLAP model, the problem with the random forest model is that it is relatively weak in dealing with high-dimensional data and time series data. The random forest model is more suitable for dealing with static data and low-dimensional data. Therefore, the performance of the random forest model is poor in the prediction of shale gas saturation. The applicability of the CLAP model to predict gas saturation includes four aspects: first, the quality of logging data is good, and non-formation factors can be used for prediction after correction; second, the core positioning should be accurate; third, the parameters of sensitive variables should be selected appropriately. Only by selecting sensitive logging curves of higher importance can the prediction results be optimized. The fourth aspect is the use of a large number of training sets; the number of training sets per well should be greater than 100.
In future endeavors, we aim to enhance the CLAP model by refining its attention mechanism, which, despite effectively discerning complex nonlinear relationships, still offers potential for optimization, especially across varied shale gas contexts. Additionally, while the model has shown notable efficacy with the Sichuan Basin dataset, its broader generalization across different shale gas reservoirs necessitates evaluation using data from diverse geographical regions. Subsequent studies should investigate the CLAP model’s applicability as it extends beyond its current domain, making it a potential tool for predicting various time series data. This includes predicting EUR, TOC, and porosity in shale reservoir research, as well as applications in agriculture, soil analysis, weather forecasting, traffic data interpretation, and power load data analysis. The predictive insights garnered from the CLAP model can offer valuable technical guidance for expert decision-making across these sectors.

4. Conclusions

Traditional logging evaluation methods, often based on empirical relationships or statistical regression, struggle to delineate complex nonlinear relationships within logging data. In response, we introduce CLAP, a model that harnesses deep neural networks, particularly integrating convolutional neural networks (CNNs) and Long Short-Term Memory (LSTM) with an attention mechanism, to discern and extract concealed features from extensive logging data. This innovative approach notably enhances prediction accuracy, yielding outcomes of substantial practical value.
The high accuracy of the CLAP prediction model can be attributed to two primary factors: the introduction of the attention mechanism and the integration of CNN and LSTM neural network structures. The attention mechanism assigns differential weights to segments of input data, effectively emphasizing relevant details and diminishing irrelevant information. This prioritization, facilitated by the attention mechanism, empowers the model to hone in on the most informative features, thereby augmenting its predictive capabilities. In CLAP, the CNN captures local features from gas saturation data, while the LSTM models long-term sequence dependencies. Their combined synergy enhances the accuracy of gas saturation predictions.
Nine logging curves from 27 shale gas wells in the Changning area of the Southern Sichuan Basin were employed to train the CLAP model. This model was subsequently applied to predict gas saturation in five additional shale gas wells not included in the training phase. When compared to both the Archie model and the random forest model, the CLAP model showcased superior accuracy in predicting shale gas saturation. Notably, the CLAP model achieved an R2 score of 0.762 and a mean square error (MSE) score of 0.934, underscoring its outstanding statistical performance and potential utility in gas saturation prediction.

Author Contributions

Conceptualization, T.Z. and X.Y.; methodology, C.Z. and S.Z.; software, T.Z. and S.Z.; validation, T.Z.; formal analysis, D.Z. and Z.S.; investigation, X.Y. and C.Z.; resources, T.Z.; data curation, D.Z.; writing—original draft preparation, T.Z. and R.J.; writing—review and editing, S.L. and Y.Z.; supervision, R.J. and Y.Z.; project administration, S.L. and G.W.; funding acquisition, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data is unavailable due to privacy or ethical restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Burnaman, M.D.; Xia, W.W.; Shelton, J. Shale gas play screening and evaluation criteria. China Pet. Explor. 2009, 14, 51–64. [Google Scholar]
  2. Kaleris, V.K.; Ziogas, A.I. Using electrical resistivity logs and short duration pumping tests to estimate hydraulic conductivity profiles. J. Hydrol. 2020, 590, 125277. [Google Scholar] [CrossRef]
  3. Simandoux, P. Dielectric Measurements on Porous Media Application to the Measurement of Water Saturations: Study of the Behaviour of Argillaceous Formations; Supplementary Issue; Institut Francais du Petrole: Rueil-Malmaison, France, 1963; Volume 18, pp. 193–215. [Google Scholar]
  4. Mahdi, Z.A.; Farman, G.M. A Review on Models for Evaluating Rock Petrophysical Properties. Iraqi J. Chem. Pet. Eng. 2023, 24, 125–136. [Google Scholar] [CrossRef]
  5. Duan, X.; Wu, Y.; Jiang, Z.; Hu, Z.; Tang, X.; Zhang, Y.; Wang, X.; Chen, W. A New Method for Predicting the Gas Content of Low-Resistivity Shale: A Case Study of Longmaxi Shale in Southern Sichuan Basin, China. Energies 2023, 16, 6169. [Google Scholar] [CrossRef]
  6. Clavier, C.; Coates, G.; Dumanoir, J. Theoretical and Experimental Bases for the Dual-Water Model for Interpretation of Shaly Sands. Soc. Pet. Eng. J. 1984, 24, 153–168. [Google Scholar] [CrossRef]
  7. Schlumberger. Schlumberger Log Interpretation Principles/Applications, 8th ed.; Schlumberger: Sugar Land, TX, USA, 1989. [Google Scholar]
  8. Waxman, M.H.; Thomas, E.C. Electrical conductivities in shaly sands–II. The temperature coefficient of electrical conductivity. J. Petrol. Technol. Trans. AIME 1974, 257, 218. [Google Scholar]
  9. Waxman, M.H.; Smits, L.J.M. Electrical conductivities in oil-bearing shaly sands. Soc. Petrol. Eng. J. 1968, 8, 107–122. [Google Scholar] [CrossRef]
  10. Archie, G.E. The electrical resistivity log as an aid in determining some reservoir characteristics. Trans. AIME 1942, 146, 54–62. [Google Scholar] [CrossRef]
  11. Yong, S.H.; Zhang, C.M. Logging Data Processing and Comprehensive Interpretation; China University of Petroleum Press: Qingdao, China, 2007. [Google Scholar]
  12. Zhang, J.Y.; Sun, J.M. Log evaluation on shale hydrocarbon reservoir. Well Logging Technol. 2012, 36, 146–153. [Google Scholar]
  13. Yan, L.; Zhou, W.; Fan, J.Y.; Wu, J.; Wang, X. Log evaluation method for gas content of deep shale gas reservoirs in southern Sichuan basin, 02. Well Logging Technol. 2019, 43, 149–154. (In Chinese) [Google Scholar]
  14. Shi, W.R.; Zhang, C.M.; Zhang, Z.S.; Xiao, S.H.; Shi, Y.H.; Ren, Y. Log evaluation of gas content from Jiaoshiba shale gas reservoir in fuling gas field. Well Logging Technol. 2015, 39, 357–362. (In Chinese) [Google Scholar]
  15. Li, J.; Wu, Q.Z.; Jin, W.J.; Jing, L.; Nan, Z.Y. Logging evaluation of free-gas saturation and volume content in Wufeng-Longmaxi organic-rich shales in the Upper Yangtze Platform. China Mar. Petrol. Geol. 2019, 100, 530–539. [Google Scholar] [CrossRef]
  16. Li, P.J.; Chen, A.Q.; Fu, Y.L.; Han, W.S. Study on determing Nuclear Magnetic Resonance (NMR) T2 cutoff combined with the oil displacing water experiments. Prog. Geophys. 2019, 34, 1050–1054. (In Chinese) [Google Scholar]
  17. Aifa, T.; Baouche, R.; Baddari, K. Neuro-fuzzy system to predict permeability and porosity from well log data: A case study of Hassi R’Mel gas field, Algeria. J. Pet. Sci. Eng. 2014, 123, 217–229. [Google Scholar] [CrossRef]
  18. Al-Mudhafar, W.J. Integrating well log interpretations for lithofacies classification and permeability modeling through advanced machine learning algorithms. J. Pet. Explor. Prod. Technol. 2017, 7, 1023–1033. [Google Scholar] [CrossRef]
  19. Wood, D.A. Prediction of gas saturation in shale gas reservoirs: A review. J. Pet. Sci. Eng. 2020, 178, 106587. [Google Scholar] [CrossRef]
  20. Almpanis, A.; Tsourlos, P.; Vargemezis, G.; Papazachos, C. Application of crosshole electrical resistivity tomography measurements under the influence of horizontally slotted plastic cased boreholes. Near Surf. Geophys. 2022, 20, 46–63. [Google Scholar] [CrossRef]
  21. Xu, S.L.; Zhang, Y.; Li, H.; Zhang, Z.H.; Luo, Q.G.; Li, W.Y. Quantitative prediction of gas saturation in low permeability tight sandstone reservoirs with multiple parameters. Nat. Gas Explor. Dev. 2020, 45, 92–98. [Google Scholar]
  22. Huang, L.S.; Yan, J.P.; Guo, W.; Zheng, M.J.; Zhong, G.H.; Huang, Y. Saturation evaluation of low resistivity shale gas reservoir based on random forest regression algorithm. Well Logging Technol. 2023, 47, 22–28. [Google Scholar]
  23. Nembrini, S. On what to permute in test-based approaches for variable importance measures in Random Forests. Bioinformatics 2019, 15, 2701–2705. [Google Scholar] [CrossRef]
  24. Li, W. Research on Loan Overdue Prediction Method Based on LSTM-CNN. Inf. Technol. Inform. 2020, 248, 40–41. [Google Scholar]
  25. Wei, J.; Zhao, H.; Liu, D.; Jia, H.; Wang, X.; Zhang, H.; Liu, N. CNN-LSTM Short-Term Power Load Forecasting Method Based on Attention Mechanism. J. North China Electr. Power Univ. Nat. Sci. Ed. 2021, 48, 42–47. [Google Scholar]
  26. Liu, Y.Y. Air quality index prediction based on CNN-LSTM and attention mechanism. Comput. Age 2022, 355, 58–60. [Google Scholar]
  27. Wang, J.; Gao, Z.; Zhu, Y.-M. Water quality prediction of the Yellow River based on CNN-LSTM model. Yellow River 2021, 43, 96–99. [Google Scholar]
  28. Ma, F.; Tu, Z.Y.; Zhu, S.T.; Xiang, M.Y.; Sun, Y.F.; Fang, Q. Research on LSTM water level prediction model based on improved attention mechanism. Jiangxi Water Resour. Sci. Technol. 2023, 49, 162–166. [Google Scholar]
  29. Mu, J.; He, H.; Li, L.; Pang, S.; Liu, C. A Hybrid Network Intrusion Detection Model Based on CNN-LSTM and Attention Mechanism; Springer: Singapore, 2022. [Google Scholar] [CrossRef]
  30. Huang Fu, X.; Qian, H.; Huang, M. Review of deep neural networks combined with Attention mechanism. Comput. Mod. 2023, 330, 40–49. [Google Scholar]
  31. Chen, X.; Zhu, K. Lithium battery health state assessment based on CNN-LSTM and attention mechanism. Ind. Control Comput. 2012, 35, 80–82. [Google Scholar]
  32. Oluwole, A.G.; Nosa, E.J. Morphological Analysis of an Organic-rich Shale: Implication for Potential Gas Energy Generation, Witbank Coalfield, South Africa. Microsc. Microanal. 2023, 29 (Suppl. S1), 766–767. [Google Scholar]
  33. Niu, W.; Sun, Y.; Yang, X.; Lu, J.; Zhao, S.; Yu, R.; Liang, P.; Zhang, J. Toward Production Forecasting for Shale Gas Wells Using Transfer Learning. Energy Fuels 2023, 37, 5130–5142. [Google Scholar] [CrossRef]
  34. Kang, D.; Ma, J.; Zhao, Y.P. Perspectives of Machine Learning Development on Kerogen Molecular Model Reconstruction and Shale Oil/Gas Exploitation. Energy Fuels 2023, 37, 98–117. [Google Scholar] [CrossRef]
  35. Saporetti, C.M.; Fonseca, D.L.; Oliveira, L.C. Hybrid machine learning models for estimating total organic carbon from mineral constituents in core samples of shale gas fields. Mar. Petrol. Geol. 2022, 143, 105783. [Google Scholar] [CrossRef]
  36. Jiang, Z.; Qi, Q.; Jiang, X.; Meng, J.; Wang, X.J. An efficient rock physics scheme for estimating crack density and fluid saturation of shale gas reservoir. Front. Earth Sci. 2022, 9, 829244. [Google Scholar] [CrossRef]
  37. Zhai, S.; Geng, S.; Li, C.; Gong, Y.; Jing, M.; Li, Y. Prediction of gas production potential based on machine learning in shale gas field: A case study. Energy Sources Part A Recovery Util. Environ. Eff. 2022, 44, 6581–6601. [Google Scholar] [CrossRef]
  38. Tang, L.; Song, Y.; Li, Q.; Pang, X.; Jiang, Z.; Li, Z. A Quantitative Evaluation of Shale Gas Content in Different Occurrence States of the Longmaxi Formation: A New Insight from Well JY-A in the Fuling Shale Gas Field, Sichuan Basin. Acta Geol. Sin.-Engl. Ed. 2019, 93, 400–419. [Google Scholar] [CrossRef]
  39. Li, Y.F.; Sun, W.; Liu, X.W.; Zhang, D.W.; Wang, Y.C.; Liu, Z.Y. Study of the relationship between fractures and highly productive shale gas zones, Longmaxi Formation, Jiaoshiba area in eastern Sichuan. Petrol. Sci. 2018, 15, 498–509. [Google Scholar] [CrossRef]
  40. Li, P.; Jiang, Z.; Zheng, M.; Bi, H.; Chen, L. Estimation of shale gas adsorption capacity of the Longmaxi Formation in the Upper Yangtze Platform, China. J. Nat. Gas Sci. Eng. 2016, 34, 1034–1043. [Google Scholar] [CrossRef]
  41. Zhao, Z.; Su, S.; Shan, X.; Li, X.; Zhang, J.; Jing, C.; Ren, H.; Li, A.; Yang, Q.; Xing, J. Lithofacies identification of shale reservoirs using a tree augmented Bayesian network: A case study of the lower Silurian Longmaxi formation in the changning block, South Sichuan basin, China. Geo Sci. Eng. 2023, 221, 211385. [Google Scholar] [CrossRef]
  42. Chen, Z.Q. Quantitative seismic prediction technique of marine shale TOC and its application: A case from the Longmaxi Shale Play in the Jiaoshiba area, Sichuan Basin. Nat. Gas Ind. 2014, 34, 24–29. [Google Scholar]
  43. Kuyumani, E.M.; Hasan, A.N.; Shongwe, T. A Hybrid Model Based on CNN-LSTM to Detect and Forecast Harmonics: A Case Study of an Eskom Substation in South Africa. Electr. Power Compon. Syst. 2023, 51, 746–760. [Google Scholar] [CrossRef]
  44. Chaudhary, M.; Gastli, M.S.; Nassar, L.; Karray, F. Deep Learning Approaches for Forecasting Strawberry Yields and Prices Using Satellite Images and Station-Based Soil Parameters. arXiv 2021, arXiv:2102.09024. [Google Scholar]
  45. Wu, P.W.; Huang, Z.H.; Pian, Y.P.; Xu, L.X.; Li, J.L.; Chen, K.C. A Combined Deep Learning Method with Attention-Based LSTM Model for Short-Term Traffic Speed Forecasting. J. Adv. Transport. 2020, 2020, 8863724. [Google Scholar] [CrossRef]
  46. Ai, X.; Li, S.; Xu, H. Short-term wind speed forecasting based on two-stage preprocessing method, sparrow search algorithm and long short-term memory neural network. Energy Rep. 2022, 8, 14997–15010. [Google Scholar] [CrossRef]
  47. Gupta, A.; Sharma, M.; Srivastava, A. Intelligent Software Bug Prediction Framework with Parameter-Tuned LSTM with Attention Mechanism Using Adaptive Target-Based Pooling Deep Features. Int. J. Reliab. Qual. Saf. Eng. 2023, 30, 2350005. [Google Scholar] [CrossRef]
  48. Shi, Z. Graph neural networks and attention-based CNN-LSTM for protein classification. arXiv 2022, arXiv:2204.09486. [Google Scholar]
  49. Lin, J.; Ma, J.; Zhu, J.; Cui, Y. Short-term load forecasting based on LSTM networks considering attention mechanism. Int. J. Electr. Power 2022, 137, 107818. [Google Scholar] [CrossRef]
  50. Kota, V.R.; Munisamy, S.D. High accuracy offering attention mechanisms based deep learning approach using CNN/bi-LSTM for sentiment analysis. Int. J. Intell. Comput. 2022, 15, 61–74. [Google Scholar] [CrossRef]
  51. Zhang, D.; Wang, S. A protein succinylation sites prediction method based on the hybrid architecture of LSTM network and CNN. J. Bioinf. Comput. Biol. 2022, 20, 2250003. [Google Scholar] [CrossRef]
  52. Liang, Y.; Lin, Y.; Lu, Q. Forecasting gold price using a novel hybrid model with ICEEMDAN and LSTM-CNN-CBAM. Expert Syst. Appl. 2022, 206, 117847. [Google Scholar] [CrossRef]
  53. Kanwal, A.; Lau, M.F.; Ng, S.P.H.; Sim, K.Y.; Chandrasekaran, S. BiCuDNNLSTM-1dCNN—A hybrid deep learning-based predictive model for stock price prediction. Expert Syst. Appl. 2022, 202, 117123. [Google Scholar] [CrossRef]
Figure 1. Correlation analysis of 10 logging curve data with gas saturation (based on R-squared parameter characterization).
Figure 1. Correlation analysis of 10 logging curve data with gas saturation (based on R-squared parameter characterization).
Processes 11 02645 g001
Figure 2. Experimental flow chart of water saturation prediction method in this study.
Figure 2. Experimental flow chart of water saturation prediction method in this study.
Processes 11 02645 g002
Figure 3. The memory cell structure of an LSTM. The dashed box encompasses the core section of the LSTM.
Figure 3. The memory cell structure of an LSTM. The dashed box encompasses the core section of the LSTM.
Processes 11 02645 g003
Figure 4. Attention structure diagram in this study.
Figure 4. Attention structure diagram in this study.
Processes 11 02645 g004
Figure 5. Attention-based CNN-LSTM model for water saturation prediction.
Figure 5. Attention-based CNN-LSTM model for water saturation prediction.
Processes 11 02645 g005
Figure 6. CLAP architecture in this study.
Figure 6. CLAP architecture in this study.
Processes 11 02645 g006
Figure 7. Loss plot for CLAP model with 1436 samples, 500 epochs. (a) Loss values vs. epoch, (b) R2 vs. Epoch.
Figure 7. Loss plot for CLAP model with 1436 samples, 500 epochs. (a) Loss values vs. epoch, (b) R2 vs. Epoch.
Processes 11 02645 g007
Figure 8. Comparison of gas saturation in Well A predicted using three models (CLAP model, random forest model, Archie model) with gas saturation obtained from experimental testing.
Figure 8. Comparison of gas saturation in Well A predicted using three models (CLAP model, random forest model, Archie model) with gas saturation obtained from experimental testing.
Processes 11 02645 g008
Figure 9. Scatter plot of gas saturation in Well A predicted by three models (CLAP model, random forest model, Archie model) versus gas saturation obtained from experiments. (a) Gas saturation calculated by CLAP model vs. experimental values of gas saturation, (b) gas saturation calculated by Archie model vs. experimental values of gas saturation, (c) Gas saturation calculated by random forest model vs. experimental values of gas saturation.
Figure 9. Scatter plot of gas saturation in Well A predicted by three models (CLAP model, random forest model, Archie model) versus gas saturation obtained from experiments. (a) Gas saturation calculated by CLAP model vs. experimental values of gas saturation, (b) gas saturation calculated by Archie model vs. experimental values of gas saturation, (c) Gas saturation calculated by random forest model vs. experimental values of gas saturation.
Processes 11 02645 g009
Table 1. Training parameters in this study.
Table 1. Training parameters in this study.
ParametersValue
Convolution kernel size3 × 3
Number of convolutional layers3
Pooling operationMax pooling
Activation functionsReLU
Table 2. CNN model structural parameters.
Table 2. CNN model structural parameters.
ParametersValue
Number of LSTM layers3
Number of hidden units256
Gated unit activation functiontanh
Gated unit activation functionSigmoid
Table 3. Structural parameters of LSTM models.
Table 3. Structural parameters of LSTM models.
ParametersValue
Number of hidden units in fully connected layers128
Weight calculation method5000
Table 4. ATT parameters.
Table 4. ATT parameters.
ParametersValue
Learning rate0.0002
Training batch5000
Number of samples64
Table 5. Evaluation indicators of model results.
Table 5. Evaluation indicators of model results.
Evaluation MetricsNumerical Values
R2 score0.762
Mean square error0.934
Mean absolute error0.735
Table 6. Comparison of the prediction results of three models (CLAP model, random forest model and Archie model) in Well A and gas saturation data from the experiments.
Table 6. Comparison of the prediction results of three models (CLAP model, random forest model and Archie model) in Well A and gas saturation data from the experiments.
DepthSg_CoreCLAP ModelArchie ModelRandom Forest ModelDepthSg_CoreCLAP ModelArchie ModelRandom Forest Model
Sg_ATTRelative ErrorSg_ArchieRelative ErrorSg_RFRelative ErrorSg_ATTRelative ErrorSg_ArchieRelative ErrorSg_RFRelative Error
1259.5957.1855.771.4142.0715.1145.2911.891276.6651.6051.200.4016.3435.2620.9730.63
1261.4750.4047.652.7533.2317.1754.994.591277.7241.2344.503.2724.4616.7729.8811.35
1262.0957.1857.100.0845.4311.7545.2511.931278.5056.3153.632.6834.4321.8847.858.46
1263.032.011.230.784.176.1814.5912.581279.5145.4946.611.1232.8012.6944.600.89
1263.9750.4050.640.2435.1415.2639.7610.641280.0450.4148.991.4234.8215.5944.226.19
1265.1431.0232.331.319.5921.4318.5912.431280.5152.2554.021.7741.1711.0856.564.31
1265.9654.8153.281.5338.0516.7622.4632.351281.5253.5651.042.5222.3431.2246.726.84
1266.4952.3852.300.0825.3527.0446.066.321282.0145.4943.541.9528.9016.5942.023.47
1267.4847.0246.660.3632.3414.6835.0711.951282.5232.9929.203.7920.3012.6936.913.92
1267.9150.6049.650.9536.6014.0041.589.021283.0152.2551.950.3040.6511.6057.385.13
1268.1828.8027.830.9711.3617.4420.897.911283.5246.7243.413.3129.3217.4077.6230.90
1268.9952.3849.572.8137.8214.5654.982.601284.5048.1346.181.9530.7817.3559.9411.81
1269.2050.9949.271.7230.9220.0746.224.771285.0232.9934.321.3329.153.8435.152.16
1270.6828.8028.380.4222.346.4617.6511.151285.5647.5845.661.9243.923.6656.709.12
1270.7619.4416.033.416.9312.5115.354.091286.6849.7849.900.1246.393.3955.625.84
1271.0049.8448.151.6923.3426.5062.6712.831287.0048.1346.012.1231.1616.9740.167.97
1272.1249.3246.482.8434.1715.1524.8124.511288.3145.2845.340.0632.7612.5262.5717.29
1273.1758.0858.670.5932.3225.7649.558.531288.6097.4396.261.1781.5215.9186.2811.15
1273.2619.4418.830.616.9412.5012.806.641289.1849.7846.792.9938.1811.6046.623.16
1274.1651.6052.130.5338.9812.6226.7224.881290.5859.8657.881.9822.1937.6764.614.75
1275.2241.2341.880.6522.8018.4312.2528.981290.8145.2846.691.412.0143.2778.4733.19
1275.6758.0857.530.5555.582.5027.7930.291291.1097.4395.461.9759.9037.5351.4445.99
1276.5140.1039.041.0618.7921.3117.7522.351292.0057.7954.822.9743.3714.4266.608.81
1293.0859.8655.754.1145.6314.2363.393.531306.5976.5577.470.9264.1612.3971.345.21
1293.1149.4747.531.9432.0917.3871.9922.521307.0472.6974.181.4958.5214.1782.189.49
1294.4651.0553.132.0832.4618.5961.1110.061308.0270.4370.030.4056.3514.0884.1113.68
1295.4318.7317.301.433.0015.7329.6610.931308.6270.2770.150.1258.0312.2482.8812.61
1295.6149.4749.980.5133.4016.0741.737.741309.0976.5577.520.9763.5513.0085.348.79
1295.8155.4857.652.1754.830.6544.0811.401309.6483.8083.210.5968.3415.4694.4910.69
1296.4653.1356.523.3948.264.8743.179.961310.0976.5278.511.9958.1018.4268.677.85
1296.9651.0553.482.4345.475.5839.2211.831310.7472.3869.233.1556.5915.7965.387.00
1297.9318.7316.911.8211.926.8113.325.411311.1270.2767.452.8256.9913.2859.3810.89
1298.3155.4855.510.0348.536.9544.7710.711311.6178.2974.863.4363.6914.6075.392.90
1298.4444.4242.761.6639.514.9147.192.771312.3874.9271.883.0465.849.0865.659.27
1298.9653.1354.751.6238.1614.9775.5822.451312.5146.6943.702.9936.6010.0942.234.46
1299.4751.2450.141.1033.9417.3058.036.791313.2472.3873.611.2357.0315.3582.4910.11
1300.4941.5640.830.7327.6913.8753.1411.581313.5667.7365.522.2151.4816.2568.240.51
1300.6052.2151.570.6435.3316.8843.288.931314.1178.2976.621.6759.2819.0192.3914.10
1301.4248.1247.280.8433.6014.5239.448.681314.8874.9277.612.6969.465.4682.397.47
1301.9751.2448.642.6033.5417.7041.0810.161315.0146.6945.700.9934.6512.0451.945.25
1302.4454.1752.721.4553.141.0347.206.971316.0667.7366.731.0052.8514.8856.8810.85
1303.8765.4063.372.0359.845.5660.305.101316.4573.8472.311.5362.1211.7264.079.77
1303.9248.1245.083.0423.1624.9649.121.001317.6457.1357.080.0539.3517.7852.035.10
1304.5472.6972.500.1947.4825.2175.372.681318.1469.3969.030.3653.7815.6162.177.22
1305.5270.4368.911.5243.4626.9785.1314.701319.3462.6964.411.7248.0814.6162.740.05
1306.3765.4067.291.8963.651.7558.486.921320.1457.1356.071.0656.240.8967.4710.34
Table 7. Comparison of the differences between the calculation results from three saturation models and core analysis in 5 wells.
Table 7. Comparison of the differences between the calculation results from three saturation models and core analysis in 5 wells.
Well NameNumber of Experimental SamplesCLAP ModelArchie ModelRandom Forest Model
R2RMSEMAER2RMSEMAER2RMSEMAE
Well A920.971.871.570.7816.9314.960.5713.4610.67
Well B1220.981.021.520.7212.3417.340.4517.458.45
Well C1320.960.891.440.7613.2612.340.6719.2212.66
Well D620.841.631.880.4615.7425.340.4322.4216.45
Well E780.741.722.010.5514.6627.430.2627.4713.24
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Zhang, C.; Zhao, S.; Zhou, T.; Zhang, D.; Shi, Z.; Liu, S.; Jiang, R.; Yin, M.; Wang, G.; et al. CLAP: Gas Saturation Prediction in Shale Gas Reservoir Using a Cascaded Convolutional Neural Network–Long Short-Term Memory Model with Attention Mechanism. Processes 2023, 11, 2645. https://0-doi-org.brum.beds.ac.uk/10.3390/pr11092645

AMA Style

Yang X, Zhang C, Zhao S, Zhou T, Zhang D, Shi Z, Liu S, Jiang R, Yin M, Wang G, et al. CLAP: Gas Saturation Prediction in Shale Gas Reservoir Using a Cascaded Convolutional Neural Network–Long Short-Term Memory Model with Attention Mechanism. Processes. 2023; 11(9):2645. https://0-doi-org.brum.beds.ac.uk/10.3390/pr11092645

Chicago/Turabian Style

Yang, Xuefeng, Chenglin Zhang, Shengxian Zhao, Tianqi Zhou, Deliang Zhang, Zhensheng Shi, Shaojun Liu, Rui Jiang, Meixuan Yin, Gaoxiang Wang, and et al. 2023. "CLAP: Gas Saturation Prediction in Shale Gas Reservoir Using a Cascaded Convolutional Neural Network–Long Short-Term Memory Model with Attention Mechanism" Processes 11, no. 9: 2645. https://0-doi-org.brum.beds.ac.uk/10.3390/pr11092645

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop