Introduction

The best definition of acoustic impedance is given in medical physics (Suzuki et al. 2019). They defined it as the resistance to the propagation of ultrasound waves through tissues. In parallel, the Earth’s layers can be likened to these biological tissues. The acoustic impedance is derived within each layer by multiplying the density of the material by its acoustic velocity (g/cm3 × m/s). Based on the lithology of the layer, it is understood that porosity is typically high in sand and low in shale. An increase in acoustic impedance implies a decrease in porosity (Agbadze et al. 2022). Since low acoustic impedance and high porosity are conducive to accumulating additional hydrocarbons in each layer, sands are more likely than shales to accumulate hydrocarbons (Ali and Al-Shuhail 2018). Recognizing this, there is a growing emphasis on developing methods to determine acoustic impedance as an intrinsic property of rock layers. Current research methodologies for estimating acoustic impedance can be broadly categorized into nonmachine learning-based and machine learning (ML)-based techniques. Some examples of nonmachine learning-based methods include direct, iterative, and nonlinear inversion methods (Liu et al. 2018).

A review of the latest ML techniques for acoustic impedance estimation was conducted by Zeng et al. (2021). They introduced new ML methods to predict reservoir parameters using both post- and prestack seismic attributes. Hampson et al. (2001) utilized multiattribute transforms and neural networks to predict log characteristics from seismic data. Cracknell and Reading (2013) analyzed aircraft and satellite data using random forest (RF) and support vector machines to identify lithologic contact zones. Harris and Grunsky (2015) employed geophysical and geochemical data to predict lithology using RF. Zhang et al. (2018) conducted an experiment using deep neural networks and convolutional neural networks (CNNs) to predict seismic lithology. Biswas et al. (2019) and Das and Mukerji (2020) advocated for pre- and poststack seismic inversion using CNNs. Priezzhev et al. (2019) compared various machine learning-supported regression models, including random forest, nearest neighbor, neural network, and adaptive classifier-ensemble models. According to a recent review by Zeng et al. (2021), the random forest method is deemed one of the most effective methods for addressing highly nonlinear problems. Our research aims to explore the methodology proposed by Mardani, which employs multilayer neural backpropagation to estimate acoustic impedance using six distinct seismic attribute inputs: amplitude, second derivatives, trace gradient, quadrature amplitude, instantaneous frequency, and gradient magnitude (Mardani and Thrust 2020).

This study aims to estimate acoustic impedance with enhanced well log resolution using various machine learning methods. These methods leverage extended inputs of seismic attributes to achieve greater accuracy. Several techniques can assess the accuracy of our predictions, with one notable approach being the coefficient of determination (R2) values. This research aims to achieve R2 values near one, signifying that our predictions closely align with the actual acoustic impedance values. To meet these objectives, we face several challenges: How can we determine the acoustic impedance on a continuous well log scale using seismic attribute inputs that differ from those reported earlier in the literature? For instance, how does the extra trees regressor algorithm compare to the random forest and the multilayer perceptron regression algorithm? Moreover, how can we effectively apply these methods to real-world data for interpretation?

Methodology

This study primarily explored the application of machine learning in predicting acoustic impedance, contrasting it with the conventional band-limited impedance (BLIMP) inversion method. The groundwork data for the traditional and ML approaches were described in a prior paper (Mardani and Thrust 2020). In our study, we utilized BLIMP acoustic impedance solely for comparative analysis.

The implementation of acoustic impedance based on recursive inversion using poststack time-migrated seismic data and well logs results in a band-limited inversion. This inversion has a surface seismic frequency ranging from approximately 10–50 Hz (Mardani and Thrust 2020; Russell 1988).

Machine learning methods

Machine learning methods are utilized through the scikit-learn 1.2.1 library (Pedregosa et al. 2011). In regard to training and prediction, the input consists of depth, two-way travel time, and various seismic attributes. These attributes encompass amplitudes, their integrals, trace gradients, quadrature amplitudes, second derivatives, gradient magnitudes, instantaneous frequencies, phases, and cosine phases. The primary target for these procedures is the well log acoustic impedance, commonly referred to as the true AI. The output generated is the predicted acoustic impedance at the well log resolution.

Algorithms

In this study, three models were employed: multilayer perceptron regression (MLPR), random forest regression (RFR), and extra tree regression (ETR). A brief description of each is given below.

Multilayer perceptron regression (MLPR)

The multilayer perceptron (MLP) is encapsulated within the MLPRegressor class, which employs a backpropagation-trained multilayer perceptron. Given that the output consists of a series of continuous numbers, the square error serves as the loss function (Pedregosa et al. 2011). Broadly speaking, the inherent specifications of the MLPR function include a hidden layer size of 100, the rectified linear unit (ReLU) activation function (which outputs the input directly if it is positive and zero if it is negative), the Adam optimizer for optimization, an autobatch size, a constant learning rate, and a maximum iteration count of 1000, among others.

(1)
Random forest regression (RFR)

A random forest operates as a meta-estimator aggregating numerous decision trees, each trained on various subsamples of datasets. This aggregation aims to enhance the prediction accuracy and curtail overfitting (Pedregosa et al. 2011). The built-in function encompasses several estimators set at 100 and employs the squared error as its criterion, among other features.

$$\hat{y} = \frac{1}{B}\sum\limits_{b = 1}^{B} {f_{b} \left( x \right)}$$
(2)

Step-by-step mathematical derivation of random forest regression using scikit-learn (Pedregosa et al. 2011):

  • Initialize max_depth as the maximum depth of each tree and n_estimators as the number of trees in the forest.

  • For every forest tree, create a new dataset by randomly selecting a portion of the training data (with replacement), choose a subset of the features at random to consider while dividing each tree node, and build a decision tree using the new dataset and selected characteristics, with a maximum depth of max_depth.

  • To predict a new data point, each tree in the forest was predicted, and the predictions were averaged to obtain the final prediction.

The mathematical formulation for the random forest algorithm, as outlined by Pedregosa et al. (2011), is as follows:

  1. 1.

    Initialization

  • n_estimators: This represents the number of trees in the forest.

  • max_depth: This denotes the maximum depth of each decision tree.

  1. 2.

    For each tree in the forest

  2. a.

    Dataset creation

  • Let D be the original dataset containing N samples.

  • D′, a new dataset of size N′ (where N′ ≤ N), is constructed by randomly selecting N′ samples from D with replacement.

  1. b.

    Feature Selection.

  • Let n be the total number of features in the dataset.

  • Here, n′ is defined as the number of features to consider when splitting each node in the tree, ensuring that n′ ≤ n.

  • Let F represent the set of all features in the dataset.

  • F′, a new set of features, is constructed by randomly selecting n′ features from F without replacement.

  1. c.

    Decision Tree Construction.

  • A decision tree, denoted as T, is built using dataset D′ and feature set F′, ensuring that it does not exceed the specified maximum depth.

  1. 4.

    Prediction for a New Data Point:

  2. 5a.

    Individual Tree Predictions:

  • Let T1, T2…., and Tn be the n decision trees in the forest.

  • For a new data point x to be predicted, let yi be the prediction made by tree Ti for x.

  1. b.

    Final Prediction:

    The final prediction y for x is determined by averaging the predictions y1, y2...., yn.

Extra tree regressor

For implementation in this class, averaging is employed as part of the meta-estimator approach. This method fits multiple randomized decision trees, often referred to as "extra trees," on various subsamples of the dataset. This strategy aims to enhance the prediction accuracy and mitigate overfitting (Pedregosa et al. 2011) (Figs. 1, 2, 3, 4, 5).

Fig. 1
figure 1

Comparison between BLIMP (nonmachine learning) and machine learning methods for estimating the acoustic impedance (Maurya and Singh 2018, 2019; Mardani and Thrust 2020)

Fig. 2
figure 2

Inputs and outputs for the ML models

Fig. 3
figure 3

Graphical illustration of the multilayer perceptron regression algorithm (Petre 2021)

Fig. 4
figure 4

Graphical illustration of the random forest regression algorithm

Fig. 5
figure 5

Graphical illustration of the extra tree regressor algorithm (Chu et al. 2021; Geurts et al. 2006)

Results and discussion

Data visualization and analysis

We utilized the dataset from Mardani and Thrust (2020) to illustrate our selected method. The data are depicted in Fig. 6, which comprises an extension to nine inputs of seismic attributes in addition to those of depth and TWTT. The details and definitions of each attribute can be found in Appendix A. A plot of the zero-offset seismic trace is shown in the first track on the left. Seismic traces for all wells are used to extract the following training input attributes: the amplitude of the seismic trace, integral, second derivative, quadrature amplitude, trace gradient, gradient magnitude, instantaneous frequency, phase, and cosine of the phase as the nine of eleven input features, while the acoustic impedance at well log resolution is the target (Mardani and Thrust 2020).

Fig. 6
figure 6

This figure shows a single well location with a zero-offset seismic trace in the first track on the left. Seismic traces for all wells are used to extract the following training input attributes: the amplitude of the seismic trace, integral, second derivative, quadrature amplitude, trace gradient, gradient magnitude, instantaneous frequency, phase, and cosine of the phase as the input features, while the acoustic impedance at well log resolution (AI) is the target (Mardani and Thrust 2020)

The relationships between the attributes are more clearly visualized in the cross-correlation matrix presented in Fig. 7. Within this matrix, values closer to 1 indicate a positive correlation, while those nearing -1 signify a negative correlation. Leveraging the data splitting feature, we designate the AI log (acoustic impedance derived from the well log) as both the target and output. Other than AI_HRS_inv are treated as inputs. The AI_HRS_inv represents the band-limited inversion from prior research, generated using Hampson Russel software (Mardani 2020). According to the matrix, the most robust input‒output relationship exists between Quadr and AI_Log.

Fig. 7
figure 7

The cross-correlation matrix was computed for the input features, including the seismic attributes, depth and TWTT, while the AI log was the target variable. The results revealed positive correlations among all the input features, except for the second derivative, trace gradient, and frequency, which exhibited negative correlations. Despite their negative correlations, these three features were retained for evaluation, aiming to compare their performance with that of the previously referenced model. In relation to the AI log, the highest positive correlation was observed with the quadrature trace, while the lowest negative correlation was associated with the trace gradient

To evaluate the performance of the selected methods, we present the errors in the prediction results for RFR, MLPR, and ETR using the default parameters in Table 1. The comparison reveals that ETR consistently exhibits lower values for MAE, MSE, and RMSE than the other methods. To illustrate the performance of the three selected methods, we compare the R2 values between the actual and predicted impedance values in Fig. 8. Notably, the R2 value for the ETR surpasses those for both the RFR and MLPR. Specifically, ETR achieves the highest R2, followed by RFR, outperforming MLPR. These findings further reveal that, in comparison to Mardani and Thrust (2020), who achieved an R2 of 0.88 using the TensorFlow platform, both our ETR and RFR methods yield superior R2 values, while our MLPR produces a slightly inferior result. Such discrepancies can arise due to variations in parameters such as the number of features, and platforms including different libraries utilized. In our study, we employed the sci-kit Learn 1.2.1 libraries for computations, whereas Mardani and Thrust (2020) utilized TensorFlow and Keras. Furthermore, we compare histograms of the prediction errors of the selected methods in Fig. 9. It is evident that the prediction error of the ETR method primarily spikes at zero error, unlike that of the RFR and MLPR methods, which display a broader spread of nonzero errors. This observation leads to the ranking of prediction error quality for the three-based methods as ETR > RFR > MLPR.

Table 1 Error values in prediction using the three methods
Fig. 8
figure 8

R2 (determination coefficient) values depicting the correlation between actual and predicted acoustic impedance from multilayer perceptron regression (MLPR), random forest regression (RFR), and extra tree regression (ETR). The results reveal that the predictions from the ETR model closely align with the real data, indicating superior performance in capturing the relationship between the predicted and actual acoustic impedance

Fig. 9
figure 9

Prediction errors for the MLPR, RFR, and ETR methods. This figure employs error bars to convey the variability in prediction accuracy. Notably, the ETR method exhibits smaller error bars, signifying a reduced margin of error compared to both the MLPR and RFR methods. This underscores the superior predictive performance of the ETR model in these data analyses

The learning characteristics of the three regression methods can be discerned from Fig. 10. The training and testing scores of the RFR method gravitate toward 1, whereas the MLPR hovers approximately 0.6, although it continues to rise with an increase in training size. This suggests that the RFR method requires only a minimal training size, as it plateaus with a flat curve after reaching a training size of approximately 4000. This indicates that, given the same training size, the learning efficacy of RFR surpasses that of the MLPR method. Notably, the ETR, which stands out as our most effective method, consistently delivers outstanding training outcomes, ranging from a training size of 500 and maintaining this performance consistently to the endpoint, akin to the other two methods, ultimately yielding the most commendable test scores.

Fig. 10
figure 10

The learning curves of three regression models—multilayer perceptron regression (MLPR), random forest regression (RFR), and extra tree regression (ETR)—are shown. Notably, for the same amount of training data, the learning curve of ETR stands out, demonstrating superior performance with the best training and testing scores among the three models

The final estimations of the acoustic impedance are shown in Fig. 11. The band-limited AI was reproduced using available data from Mardani and Thrust (2020) to investigate how the property changes with varying depth. They generally follow a similar trend but are unable to match the same variation for localized parts (Mardani and Thrust 2020). This figure juxtaposes the acoustic impedance derived from well logs, the predicted AI using the extended inputs, and the band-limited inversion. From the figure, it is evident that the AI predicted by the RFR method, represented by a red line, aligns more closely with the true AI, depicted by a blue line, than the AI predicted by the MLPR method, which is indicated by a yellow stripe. Overall, the most accurate prediction is rendered by the ETR method, highlighted by the green line. Furthermore, the predicted acoustic impedance reveals insights suggesting that hydrocarbons are likely to accumulate at depths ranging from approximately 3100–3200 m and 3300–3400 m.

Fig. 11
figure 11

This figure presents the estimated acoustic impedance derived from various methodologies, contrasting the real target AI with predictions from RFR, MLPR, and ETR. Notably, ETR-predicted AI exhibits the highest agreement with the true values. Additionally, band-limited AI, recreated using data from prior research, explores property changes across different depths. While generally following a similar trend, it falls short of reproducing identical variations in localized sections (Mardani & Thrust, 2020)

Conclusions

  1. 1.

    The estimation of acoustic impedance, utilizing extended inputs of depth, two-way travel time and seismic attributes, employed regression methods, including multilayer perceptron regression, random forest regression, and extra tree regression with the scikit-learn library.

  2. 2.

    Our RFR and ETR studies exhibited improved determination coefficient values, surpassing those reported in the literature, even with a larger dataset due to additional features.

  3. 3.

    Among the tested models, the extra tree regression model demonstrated superior performance in terms of the coefficient of determination and was particularly suitable for highly nonlinear well log scales.

  4. 4.

    We conclude that for hydrocarbon exploration based on the available acoustic impedance data, the ETR model with default parameters represents the optimal choice for more accurate predictions. Other datasets may have different options.

  5. 5.

    Future studies should consider adopting the ETR method with diverse real datasets, incorporating additional seismic attributes, and exploring alternative machine learning models. This comprehensive approach aims to identify an even more accurate machine learning model that can then be validated using real-world data.