ARTICLE IN PRESS Computers & Geosciences 35 (2009) 1933–1939 Contents lists available at ScienceDirect Computers & Geosciences journal homepage: www.elsevier.com/locate/cageo Estimation of formation strength index of aquifer from neural networks Bieng-Zih Hsieh, Chih-Wen Wang, Zsay-Shing Lin Department of Resources Engineering, National Cheng Kung University, No. 1, University Road, Tainan 701, Taiwan a r t i c l e in fo abstract Article history: Received 5 May 2008 Received in revised form 23 September 2008 Accepted 25 November 2008 The purpose of this study is to construct a model that predicts an aquifer’s formation strength index (the ratio of shear modulus and bulk compressibility, G/Cb) from geophysical well logs by using a backpropagation neural network (BPNN). The BPNN model of an aquifer’s formation strength index is developed using a set of well logging data. The model is a [4-5-1] three-layer BPNN with a four-neuron input layer (depth, gamma-ray log data, formation density log data, and sonic log data, respectively), a five-neuron hidden layer, and a one-neuron output layer (formation strength index). The optimal learning rate and momentum constant used in the BPNN model are obtained from serial combinative experiments. The inside test and outside test are implemented to check the performance of network learning and the prediction ability of the network, respectively. The results of the inside test, based on 84 training data sets from a total of 105 data sets, show that the network has been well-trained because the mean square error between the network output value and the target value from the inside test is very small (1.1 104). The results of the outside test, based on 21 testing data sets from 105 data sets, show the excellent prediction ability of the BPNN model, because the network prediction values closely track with the target values (the mean square error is 2.1 104). & 2009 Elsevier Ltd. All rights reserved. Keywords: Back-propagation neural networks Geophysical well logs Groundwater Soft computing 1. Introduction The major parameters (or formation elastic constants) related to formation strength are shear modulus, bulk compressibility, Young’s modulus, and Poisson’s ratio. Some of these parameters are related and dependent on the other parameters (Goodman, 1989). For example, shear modulus can be expressed in terms of Young’s modulus and Poisson’s ratio. More specifically, the formation strength is proportional to shear modulus and inversely proportional to bulk compressibility. Thus, Tixier et al. (1975) used the formation strength index (G/Cb), which is the ratio of shear modulus (G) to bulk compressibility (Cb), to predict formation sanding in an oil and gas reservoir. The formation strength index (G/Cb) is not only a useful indicator of potential formation sanding (Tixier et al., 1975), but can also be used as a parameter to describe the strength of an aquifer (Hsieh et al., 2007). The formation strength index (G/Cb) can be calculated from two elastic constants, such as shear modulus (G) and bulk compressibility (Cb). Both elastic constants in turn can be obtained from core measurement (Goodman, 1989) and/or geophysical well-logging analysis (Tixier et al., 1975; Temples Corresponding author. Tel.: +886 6 275 7575x62825; fax: +886 6 238 0421. E-mail addresses: bzhsieh@mail.ncku.edu.tw (B.-Z. Hsieh), wcw0328@ms56.hinet.net (C.-W. Wang), zsaylin@mail.ncku.edu.tw (Z.-S. Lin). 0098-3004/$ - see front matter & 2009 Elsevier Ltd. All rights reserved. doi:10.1016/j.cageo.2008.11.010 and Waddell, 1996; Hsieh et al., 2007). The formation strength index (G/Cb) used in this study is from geophysical well-logging analysis (Tixier et al., 1975; Hsieh et al., 2007) in which the formation density log (FDC) data, sonic log data, and natural gamma-ray (GR) log data were used to calculate the elastic constants and formation strength index. The formation density log data and natural gamma-ray log data are used to calculate the total porosity, effective porosity, and dispersed-shale index of the formation (Hsieh et al., 2007). Poisson’s ratio is estimated from the dispersed-shale index in the geophysical well logging analysis (Tixier et al., 1975; Hsieh et al., 2007). The shear modulus (G) and bulk compressibility (Cb), which are used to calculate the formation strength index (G/Cb), are then calculated from the formation density log data, sonic log data, and Poisson’s ratio (Tixier et al., 1975; Hsieh et al., 2007). Tixier et al. (1975) estimated the formation strength parameters (such as shear modulus, bulk compressibility, Poisson’s ratio, etc.) and formation strength index (the ratio of shear modulus (G) and bulk compressibility (Cb), G/Cb) of a hydrocarbon formation from the formation density log and the sonic log. In this study, a regional empirical equation was developed to estimate Poisson’s ratio based upon the dispersed-shale index without using the information of shear-wave transit time, which was difficult to detect and always lacking. Hsieh et al. (2007) estimated an aquifer’s formation strength parameters including shear modulus, bulk compressibility, Poisson’s ratio, Young’s modulus, and formation strength index (G/Cb) by using geophysical well ARTICLE IN PRESS 1934 B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 logs, including the natural gamma-ray log, the compensated formation density log, and the compensated borehole sonic log (BHC). In the process of estimating formation strength index by using well logs, the empirical equation of Poisson’s ratio and the dispersed-shale index must be used very carefully or be modified when the research area is different. In a shallow aquifer formation, the dispersed-shale index should be estimated from the formation density log and gamma-ray log, because a suitable value for the compaction factor (CP) required in the calculation of sonic porosity (fSV) is usually unclear (Hsieh, 1997). Thus some empirical equations are required. In this case, a neural network can be used to reduce the problem of derived empirical equations and unclear data. In a neural network, the relationship between input data and output data does not need to be known, because the neural network will find the correlation during the training (or learning) process from the known examples. In recent petroleum engineering research, the combination of the geophysical well logs and the aid of back-propagation neural network (BPNN) have been studied extensively. Rogers et al. (1992) constructed a three-layer backpropagation neural network to determine the formation lithology from geophysical well logs. The input layer of the BPNN model includes three neurons (parameters) representing gamma-ray, neutron, and formation density log data; the hidden layer includes four hidden neurons; and the output layer includes four neurons representing limestone, dolomite, shale, and sandstone. Mohaghegh et al. (1996) used a three-layer back-propagation neural network to predict porosity, permeability and fluid saturation from well log data (gamma-ray, formation density, and deep induction resistivity log data) and geological information (grain-size distribution, bulk density variations, and depositional environments). Helle et al. (2001) constructed two back-propagation neural networks to estimate porosity and permeability from well logs. The well log data including gamma-ray, resistivity, sonic, formation density, and neutron log data were used as input parameters to predict the formation porosity and permeability. Shokir (2004) used a back-propagation neural network to determine the water saturation in a lowresistivity formation. The BPNN model had five input neurons representing spontaneous potential, gamma-ray, deep resistivity, formation density, and neutron log data in the input layer, and one output neuron representing water saturation in the output layer. Mohaghegh et al. (2004) constructed a three-layer back-propagation neural network to determine the in-situ stress of hydrocarbon reservoir from well logs. The input parameters include well log data (spontaneous potential, gamma-ray, sonic, bulk density, and resistivity log data), depth information, tectonic stress parameters, and formation lithology. Shokir et al. (2006) used a back-propagation neural network to predict reservoir permeability from conventional well log data. The ANN technique is demonstrated by applying it to one of the Saudi Arabia’s oil fields. The field is the largest offshore oil field in the world and was deposited in a fluvial dominated deltaic environment. The ANN permeability prediction model was developed using some of the core permeability and well log data from three early development wells. The BPNN model had six input neurons representing corrected log porosity, gamma-ray, deep resistivity, formation density, neutron log, and sonic log data in the input layer, one hidden layer with 19 neurons, and one output neuron representing formation permeability. Rezaee et al. (2007) used a back-propagation neural network to predict shear-wave velocity from well log data for a sandstone reservoir of the Carnarvon basin, Australia. The well log data including gamma-ray, induction resistivity, formation density, sonic, and neutron log data were used in the input layer for predicting shear-wave velocity. Since there have been few studies focusing on the hydrogeology research to combine the geophysical well logs and the aid of artificial neural networks, the purpose of this study is to construct an aquifer formation strength index prediction model from geophysical well logs by using a back-propagation neural network. 2. Back-propagation neural network As the most popular model for engineering applications, the back-propagation neural network is a multilayer perceptron including the input layer, hidden layer, and output layer (Haykin, 2007). Each layer contains some nodes and the nodes between layers are connected by weight. The nodes of the input layer (or input nodes) are used to accept the input message and the nodes of the output layer (or output nodes) are used to produce the output value. The hidden layers, between the input layer and the output layer, are used to apply the interactive operation of nodes. When the connected weights between each layer are changed, the output value will change accordingly because of the high network connectivity. The BPNN is a supervised learning model that trains the connected weights by an error-correction learning rule (Haykin, 2007). The BPNN learning consists of two steps: a forward pass and a backward pass. During the forward pass, an input vector is introduced from the input layer and propagated through the hidden layer to the output layer, producing the output value of the network. In this pass, each input is weighted. The sum of the weighted inputs and the bias is transferred by an activation function (such as a sigmoid function) to generate the output. During the backward pass, the connected weight is adjusted by an error-correction learning rule. In this manner, the network will make the output value close to the target value. The normalization of inputs can be used to accelerate the learning process (Haykin, 2007). When the error signal, which is the difference between the target value and network output value, is very small and satisfies the convergence limit, the BPNN is considered trained and can subsequently be used for prediction. 3. Development of BPNN model To build the BPNN model of an aquifer formation strength index (G/Cb), the following well logs are used: the natural gammaray, the compensated formation density (FDC), and the compensated borehole sonic logs (Fig. 1). A total of 105 data sets, collected from the depth interval of 61–222 m, including six different formations and two different lithologies, are used in this study. The formation strength index (G/Cb) used in this study is from geophysical well-logging analysis (Tixier et al., 1975; Hsieh et al., 2007). The range of values of the formation strength index is between 0.07 1012 and 0.32 1012 psi2. In general, the scale factor for the log data (measured directly by the logging device) or the log-derived data (calculated from the log data) and core/plug measurements are close to unity if the logging devices are calibrated, and the depth correction is made (Adams, 2005). For example, the scale factor for the formation density log data (bulk density records in the wellbore) and the core measurements (bulk density measurements from laboratory) are very close to unity. As a further example, the porosity of formation, the mean difference between the log-derived porosity (calculated from the formation density log, sonic log, or neutron log) and the core-measured porosity, should be less than 0.6% (Adams, 2005). The scale factor for the log-derived porosity and core/plug measured porosity are close to unity. ARTICLE IN PRESS B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 Input Layer 1935 Hidden Layer Output Layer Depth FDC G/Cb BHC GR Fig. 2. The [4-5-1] three-layer BPNN model. Fig. 1. Geophysical well logs used to develop BPNN model. In this study, the log-derived formation strength index is used by assuming the equation for the relationship between Poisson’s ratio and the dispersed-shale index, taken from literature (Anderson et al., 1973; Tixier et al., 1975; Hsieh et al., 2007), is valid. 3.1. BPNN model structure The BPNN model developed in this study includes an input layer, a hidden layer, and an output layer. The selection of input layer parameters (or well log data) is according to the relationship with the output parameter (formation strength index). Not all geological characteristics are applicable as input parameters and such inapplicable input parameters will result in incorrect estimation. The well-logging information of FDC and BHC are the most important input parameters in the BPNN model because there is a high correlation between formation strength index and formation density (from FDC) and compression-wave transit time (from BHC) (Tixier et al., 1975; Hsieh et al., 2007). Further, the well-logging information of GR is a necessary input parameter, because the formation strength will be influenced by existing clay and silt (Dewan, 1983; Crain, 1986). There is also a tendency for formation strength to increase as the depth increases and, therefore, depth should be selected to be an input parameter (Mohaghegh et al., 1996). Thus, the input layer neurons (parameters) include Depth, GR, FDC, and BHC (Fig. 2). The output layer neuron is the formation strength index (G/Cb), which is the parameter that the BPNN will predict. The nonlinear relationship between input and output parameters cannot be determined without a hidden layer, but the networks will be too complex with excessive hidden layers. In practice, the problem of function mapping can be solved using one hidden layer with sufficient hidden layer neurons and an adequate number of weights (Goda et al., 2007). In this study, one hidden layer is adopted to build the BPNN prediction model. The optimal number of hidden layer neurons is found by a cross validation method (Haykin, 2007), in which the number of hidden layer neurons, the learning rate (Z), and the momentum constant (a) are adjusted to find the optimal number of hidden layer neurons. The optimal number of hidden layer neurons is chosen based on the smallest mean square error (MSE) among the results from different number of hidden layer neurons. Three experiment runs with a momentum constant of 0.5 were conducted using different learning rates (0.1, 0.5, or 0.9). In each run, the mean square errors of target and output values were obtained from different numbers of hidden layer neurons (from 2 to 10 neurons). The results of all three experiments with learning rates of 0.1, 0.5, or 0.9 (Tables 1–3) show that the optimal number of hidden layer neurons, based on the minimum of MSE, is five. Thus, the best BPNN structure is a three-layer perception which includes four input layer neurons, five hidden layer neurons, and one output layer neuron; this is a [4-5-1] BPNN model (Fig. 2). 3.2. Learning rate and momentum constant The efficiency of the network can be improved with an optimal learning rate and optimal momentum constant setting. In general, the learning rate of 0.5 is adopted in the range of 0.1–1.0 (Haykin, 2007), because a network with a small learning rate will converge very slowly and network with a high learning rate may result in ARTICLE IN PRESS 1936 B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 Table 1 First experiment of hidden layer neurons (nHid) with [Z ¼ 0.1, a ¼ 0.5]. nHid Mean square error (MSE) Iterations 2 3 4 5 6 7 8 9 10 2000 5000 8000 10000 12000 15000 18000 20000 0.000510 0.000581 0.000545 0.000523 0.000632 0.000550 0.000622 0.000621 0.000627 0.000331 0.000334 0.000329 0.000321 0.000361 0.000337 0.000356 0.000342 0.000349 0.000299 0.000292 0.000284 0.000278 0.000306 0.000294 0.000301 0.000299 0.000302 0.000289 0.000277 0.000268 0.000263 0.000291 0.000279 0.000286 0.000283 0.000285 0.000282 0.000266 0.000258 0.000252 0.000279 0.000269 0.000276 0.000273 0.000274 0.000274 0.000256 0.000246 0.000241 0.000267 0.000258 0.000265 0.000261 0.000262 0.000267 0.000248 0.000238 0.000232 0.000258 0.000250 0.000257 0.000252 0.000252 0.000263 0.000244 0.000233 0.000227 0.000253 0.000246 0.000253 0.000248 0.000247 Table 2 Second experiment of hidden layer neurons (nHid) with [Z ¼ 0.5, a ¼ 0.5]. nHid Mean square error (MSE) iterations 2 3 4 5 6 7 8 9 10 2000 5000 8000 10000 12000 15000 18000 20000 0.000307 0.000273 0.000265 0.000268 0.000279 0.000269 0.000279 0.000269 0.000274 0.000255 0.000233 0.000224 0.000215 0.000239 0.000235 0.000242 0.000230 0.000236 0.000214 0.000210 0.000205 0.000193 0.000227 0.000219 0.000229 0.000211 0.000219 0.000195 0.000196 0.000190 0.000180 0.000221 0.000209 0.000222 0.000198 0.000210 0.000179 0.000181 0.000171 0.000163 0.000216 0.000198 0.000215 0.000178 0.000200 0.000161 0.000158 0.000149 0.000138 0.000209 0.000177 0.000203 0.000148 0.000179 0.000146 0.000142 0.000140 0.000132 0.000199 0.000155 0.000187 0.000139 0.000155 0.000138 0.000135 0.000135 0.000128 0.000183 0.000145 0.000168 0.000134 0.000147 Table 3 Third experiment of hidden layer neurons (nHid) with [Z ¼ 0.9, a ¼ 0.5]. nHid Mean square error (MSE) iterations 2 3 4 5 6 7 8 9 10 2000 5000 8000 10000 12000 15000 18000 20000 0.000296 0.000258 0.000245 0.000246 0.000257 0.000246 0.000259 0.000243 0.000253 0.000214 0.000203 0.000205 0.000192 0.000225 0.000212 0.000228 0.000195 0.000219 0.000167 0.000161 0.000173 0.000156 0.000202 0.000187 0.000203 0.000149 0.000190 0.000148 0.000145 0.000150 0.000139 0.000178 0.000164 0.000180 0.000135 0.000165 0.000136 0.000135 0.000136 0.000128 0.000151 0.000146 0.000157 0.000130 0.000143 0.000123 0.000125 0.000124 0.000120 0.000134 0.000133 0.000137 0.000121 0.000127 0.000116 0.000118 0.000116 0.000114 0.000125 0.000125 0.000128 0.000115 0.000120 0.000112 0.000114 0.000111 0.000110 0.000122 0.000121 0.000124 0.000111 0.000116 unstable convergence during the training process. With an appropriate momentum constant, the network may have a faster convergence rate and better stability. The interval of the momentum constant is between 0.0 and 1.0, and usually the value is not greater than 0.9 (Haykin, 2007). In this study, the optimal value of the learning rate (Z) and the momentum constant (a) are determined based on the MSE from several combinative experiments. The learning rates (Z) used in the experiments range from 0.01 to 0.9, and the momentum constants (a) vary from 0.0 to 0.9. The MSE for each combinative experiment is then calculated (Figs. 3–6). The results show that the convergence speed increases as the momentum constant increases for learning rates (Z) of 0.01 or 0.1 (Figs. 3 and 4). From these two experiments, with learning rates (Z) of 0.01 or 0.1, the best momentum constant of (a) 0.9 is obtained. However, a stability problem occurred in the experiment with a momentum constant (a) of 0.9 and learning rate (Z) of 0.5, because of the oscillation of the MSE between iterations 500 and 900 (Fig. 5). For the experiments with a learning rate (Z) of 0.5 or 0.9, the results show that the best momentum constant (a) is 0.5 (Figs. 5 and 6). The comparison results (Fig. 7), based on all the above experiments (Figs. 3–6), show that the convergence speed of the ARTICLE IN PRESS B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 0.016 0.015 Mean square error (MSE) 0.02 Mean square error (MSE) 1937 =0.0 =0.1 0.01 =0.5 =0.9 0.005 0.012 = 0.01, = 0.9 = 0.1, = 0.9 0.008 = 0.5, = 0.5 = 0.9, = 0.5 0.004 0 0 5000 0 5000 10000 15000 Training iterations 20000 Fig. 3. Mean square error of target and output values versus training iterations for various momentum constant (a) for case of learning rate (Z) of 0.01. Mean square error (MSE) 0.01 0.008 = 0.0 0.006 = 0.1 0.004 0.001 0 5000 = 0.9 10000 15000 Training iterations 20000 Fig. 8. Mean square error of target and output values versus iterations of BPNN model training, using optimal learning rate and momentum constant. 0 5000 10000 15000 Training iterations 20000 Fig. 4. Mean square error of target and output values versus training iterations for various momentum constant (a) for case of learning rate (Z) of 0.1. Mean square error (MSE) 0.002 0 0 network with the combinative case of (Z ¼ 0.9, a ¼ 0.5) is the fastest and the MSE value is the smallest. Therefore, in the BPNN model the optimal value of the momentum constant (a) adopted is 0.5, and the optimal value of the learning rate (Z) is 0.9. 3.3. BPNN model training 0.005 0.004 = 0.0 0.003 = 0.1 0.002 = 0.5 0.001 = 0.9 0 0 5000 10000 15000 Training iterations 20000 Fig. 5. Mean square error of target and output values versus training iterations for various momentum constant (a) for case of learning rate (Z) of 0.5. Mean square error (MSE) 0.003 = 0.5 0.002 20000 Fig. 7. Comparison of mean square error of target and output values for four combinative settings. Mean square error (MSE) 0 10000 15000 Training iterations 0.004 0.003 = 0.0 = 0.1 0.002 = 0.5 0.001 The 105 total collected well log data sets are randomly divided into 84 training data sets and 21 testing data sets, based on an 80–20% proportion (Aminian and Ameri, 2005; Haykin, 2007). The training data sets are used in BPNN training and the testing data sets are used to estimate the prediction ability of the model. All input parameters are normalized in the range of 0–1 to accelerate network learning. In training the BPNN model, the stopping criteria of network training is set at either (1) a tolerable MSE error value of 1.0 104; or (2) 20,000 training iterations. Once the stopping criteria is satisfied, the BPNN training is complete. The result of network training (Fig. 8) shows that the mean square errors drops dramatically in less than 1000 iterations, and approaches 1.1 104 after 20,000 training iterations. The convergence speed of the network is fast and the network is stable when the optimal learning rate of 0.9 and momentum constant of 0.5 are both used. 4. Results = 0.9 0 0 5000 10000 15000 Training iterations 20000 Fig. 6. Mean square error of target and output values versus training iterations for various momentum constant (a) for case of learning rate (Z) of 0.9. After the training work of the BPNN model is completed, the performance validation and the model verification are conducted by an inside test and an outside test. The inside test is used to check the performance of network learning by using all of the training data set (Haykin, 2007). The outside test, using the testing data set, is conducted to both check the prediction ability ARTICLE IN PRESS B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 4.1. Inside test The inside test of this study uses all 84 training data sets to check the learning performance of the [4-5-1] BPNN model. The comparison plot between the target value and network output value (Fig. 9) shows that the network output value is very close to the target value. The mean square error between the network output values and the target values is 1.1 104. Therefore, the network learning performance is very good. The result of inside test (Fig. 10) further confirms the quality of the network learning performance, because the cross plot of the target value and the network output value are almost on the 451 line in the scatter diagram (Fig. 10). The calculated coefficient of determination (R2) (Draper and Smith, 1998) from the 451 line (Fig. 10) is 0.987. 4.2. Outside test In the outside test, the 21 testing (non-trained) data sets are used in the [4-5-1] BPNN model to check the model’s prediction ability and to verify that there is no network over-fitting. The comparison plot between the target value and the network output value (Fig. 11) demonstrates the excellent prediction ability of the [4-5-1] BPNN model, because the network prediction values closely track with the target value. The mean square error 0.25 Formation strength index, 1012 psi2 of the network and to verify that there is no network over-fitting. In other words, the testing data sets are also the verification samples used to overcome the problem of network over-fitting. Target value 0.20 0.15 0.10 Network output value 0.05 0.00 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 Testing data set Fig. 11. Comparison between network output value and target value from outside test. 0.25 Cross plot A=T Output value (A), 1012 psi2 1938 0.20 0.15 R2=0.977 0.10 0.05 0.00 Formation strength index, 1012 psi2 0.00 0.35 0.05 0.10 0.15 0.20 0.25 Target value (T), 1012 psi2 0.30 Fig. 12. Scatter diagram of outside test. 0.25 Target value 0.20 0.15 0.10 0.05 Network output value 0.00 1 9 17 25 33 49 41 57 65 73 81 between the network output values (prediction values) and the target values is 2.1 104. The cross plot of the outside test is very close to the 451 line in the scatter diagram (Fig. 12) and the calculated coefficient of determination (R2) is 0.977, which further shows the model’s very good predictive ability. Training data set Fig. 9. Comparison between network output value and target value from inside test. The BPNN model of the aquifer’s formation strength index (G/ Cb) has been successfully completed and the prediction ability has been proven by excellent performance during inside test and outside test validation. The conclusions of this study are as follows: 0.40 Output value (A), 1012 psi2 5. Conclusions Cross plot A=T 0.30 0.20 R2=0.987 0.10 0.00 0.00 0.10 0.20 0.30 12 Target value (T), 10 psi 2 Fig. 10. Scatter diagram of inside test. 0.40 1. A BPNN model of an aquifer’s formation strength index has been developed. The BPNN model is a [4-5-1] three-layer backpropagation neural network. The input layer includes four input neurons representing depth, GR, FDC, and BHC; the optimal hidden layer includes five hidden neurons; and the output layer includes one output neuron representing the formation strength index (G/Cb). 2. The inside test shows that the network learning performance is very good, because the network output values from the 84 training data sets are very close to the target values with a mean square error of 1.1 104. The cross plot of output and target values in the inside test are very close to the 451 line with a calculated coefficient of determination of 0.987. ARTICLE IN PRESS B.-Z. Hsieh et al. / Computers & Geosciences 35 (2009) 1933–1939 3. The outside test shows that the output values of 21 testing data sets are close to the target values with a mean square error of 2.1 104. The cross plot of output and target values in the outside test are very close to the 451 line with a calculated coefficient of determination of 0.977. The predictive ability of the BPNN model for the aquifer’s formation strength index is extremely good. References Adams, S.J., 2005. Core-to-log comparison—what’s a good match. In: Proceedings 2005 Society of Petroleum Engineers Annual Technical Conference and Exhibition, Dallas, Texas, USA, pp. 3943–3949. Aminian, K., Ameri, S., 2005. Application of artificial neural networks for reservoir characterization with limited data. Journal of Petroleum Science and Engineering 49, 212–222. Anderson, R.A., Ingram, D.S., Zanier, A.M., 1973. Determining fracture gradients from well logs. Journal of Petroleum Technology 25, 1259–1268. Crain, E.R., 1986. The Log Analysis Handbook Volume 1—Quantitative Log Analysis Methods. PennWell Publishing Company, Tulsa, Oklahoma, p. 648. Dewan, J.T., 1983. Essentials of Modern Open-Hole Log Interpretation. PennWell Publishing Company, Tulsa, Oklahoma, p. 361. Draper, N.R., Smith, H., 1998. Applied Regression Analysis. Wiley-Interscience Publication, New York, NY, p. 706. Goda, H.M., Maier, H.R., Behrenbruch, P., 2007. Use of artificial intelligence techniques for predicting irreducible water saturation—Australian hydrocarbon basins. In: Proceedings 2007 Society of Petroleum Engineers Asia Pacific Oil & Gas Conference and Exhibition, Jakarta, Indonesia, pp. 787–796. Goodman, R.E., 1989. Introduction to Rock Mechanics, second ed. John Wiley & Sons, New York, NY, p. 562. 1939 Haykin, S., 2007. Neural Networks: A Comprehensive Foundation, third ed. Prentice Hall, Inc., Upper Saddle River, New Jersey, p. 842. Helle, H.B., Bhatt, A., Ursin, B., 2001. Porosity and permeability prediction from wireline logs using artificial neural networks: a north sea case study. Geophysical Prospecting 49, 431–444. Hsieh, B.Z., 1997. Estimation of aquifer’s formation strength from well-logging data. Unpublished M.Sc. Thesis, National Chang Kung University, Tainan, Taiwan, p. 134. Hsieh, B.Z., Chilingar, G.V., Lin, Z.S., 2007. Estimation of groundwater aquifer formation-strength parameters from geophysical well log: the southwestern coastal area of Yun-Lin, Taiwan. Energy Sources, Part A 29, 97–115. Mohaghegh, S., Arefi, R., Ameri, S., Aminiand, K., Nutter, R., 1996. Petroleum reservoir characterization with the aid of artificial neural networks. Journal of Petroleum Science and Engineering 16, 263–274. Mohaghegh, S.D., Popa, A., Gaskari, R., Wolhart, S., Siegfreid, R., Ameri, S., 2004. Determining in-situ stress profiles from logs. In: Proceedings of the 2004 Society of Petroleum Engineers Annual Technical Conference and Exhibition, Houston, Texas, USA, pp. 1251–1256. Rezaee, M.R., Kadkhodaie Ilkhcji, A., Barabadi, A., 2007. Prediction of shear-wave velocity from petrophysical data utilizing intelligent systems: an example from a sandstone reservoir of Carnarvon basin, Australia. Journal of Petroleum Science and Engineering 55, 201–212. Rogers, S.J., Fang, J.H., Karr, C.L., Stanley, D.A., 1992. Determination of lithology from well logs using a neural network. The American Association of Petroleum Geologists 76, 731–739. Shokir, E.M.EI-M., 2004. Prediction of the hydrocarbon saturation in low resistivity formation via artificial neural network. In: Proceedings of the 2004 Society of Petroleum Engineers Asia Pacific Conference on Integrated Modelling for Asset Management, Kuala Lumpar, Malaysia, pp. 35–40. Shokir, E.M.El-M., Alsughayer, A.A., Al-Ateeq, A., 2006. Permeability estimation from well log responses. Journal of Canadian Petroleum Technology 45, 41–46. Temples, T.J., Waddell, M.G., 1996. Application of petroleum geophysical well logging and sampling techniques for evaluating aquifer characteristics. Ground Water 34, 523–531. Tixier, M.P., Loveless, G.W., Anderson, R.A., 1975. Estimation of formation strength from mechanical-properties log. Journal of Petroleum Technology 27, 283–293.