Valenti
Valenti
A Thesis in
Geosciences
by
Master of Science
December 2009
The thesis of Joseph Christian Adam Frank Valenti was reviewed and approved* by the
following:
Charles J. Ammon
Professor of Geosciences
Thesis Advisor
Turgay Ertekin
Professor of Petroleum and Natural Gas Engineering and George E. Trimble
Chair in Earth and Mineral Sciences
Andy Nyblade
Professor
Department of Geosciences
Kate Freeman
Professor of Geosciences
Graduate Program Chair
We compare two methods of predicting well-log porosity from seismic data. The
data consist of a suite of well logs and a full stack 3D seismic survey over Auger Field in
the deepwater Gulf of Mexico. The 3D seismic is transformed into a number of attribute
volumes. These attributes are combined in a nonlinear manner, via an Artificial Neural
Network (ANN), or in a linear manner, via multilinear regression analysis, in order to
predict the target porosity logs from the available suite of field data.
A feed-forward back propagation ANN is trained using the seismic attributes as
an input set and with the porosity logs as the output set. The linear mode uses the same
training data, but derives a series of weights which when applied to the input set
minimize the differences in a least-squares sense between the target and predicted
outputs.
In order to measure the accuracy of the attribute to porosity transformation, cross-
well validation was performed. In this procedure one well is removed from the training
set and the transformation is re-derived. The accuracy of the transformation in predicting
the log from the removed well is then measured. This is done to every well in the
training set so that we may determine a reasonable expectation for the performance of the
transformation.
We see a marked improvement of the performance of the ANN over that of the
multilinear regression. These results are evident not only in the training data but more
importantly also in the testing data.
iii
TABLE OF CONTENTS
List of Figures v
Acknowledgements ix
Introduction 1
Theoretical Background 4
Seismic Attributes 15
Results 36
Conclusions 68
Biliography 72
iv
LIST OF FIGURES
Figure 1, Page 4. Architecture of a simple Neural Network with one hidden layer. In
this scenario, the input layer is O1, the hidden layer is O2, and the output layer is O3.
Each of the circles represents a neuron. The input layer has four neurons, the hidden
layer three, and the output one. This particular network would map four inputs to one
output.
Figure 2, Page 5. Graphic display of the hyperbolic tangent sigmoid transfer function.
Note how quickly the function saturates for values greater than five. To make full use of
the shape of this transfer function, the inputs are normalized to the range -1 to 1.
Figure 3, Page 18. a) A seismic trace in blue with its amplitude envelope in red. For
every trace sample there is an attribute sample. b) The unwrapped instantaneous phase
of the trace in a). c) The instantaneous frequency. d) First Derivative with time. e)
Second Derivative with time. f) The integrated Absolute Amplitude is also used as an
attribute. It may highlight low frequency vertical trends in the data.
Figure 4, Page 23. Structure Map on top of the N interval, outlining the Lobe Facies,
Incised Fill, and well penetrations. Wells used in the training set are filled with red. The
wells penetrating the Incised Fill that are used in this study include the A02BP1, the A05,
and 426STBP1. The A01, A06, A07, and A09 penetrate the overbank deposition.
Figure 5, Page 24. a) Structure map of the NM1 with a facies interpretation overlain.
The Incised Fill Facies is shown in green; it widens and thickens to the south. b) Time
slice showing lobate geometry and the basin axis channel. c) Type well logs showing
considerable change in facies character going from the Lobe Facies (A04BP1) to the
Incised Fill Facies (A02BP1).
Figure 6, Page 26. Comparison of the original, unedited porosity log from well A01 (in
blue), and the log once it has been smoothed by a running mean filter (red).
Figure 7, Page 28. Example of the interpolation used to calculate the depth at which
each seismic sample of interest occurs. This is data from the checkshot from well A01.
At this well, the Nm1 occurs at 4.580 seconds two-way-time. The depths are calculated
for every 4 ms in a 100 ms window around the sample of interest.
Figure 8, Page 29. Porosity values that have been calculated via a spline interpolation
for each of the depth points corresponding to Figure 7.
Figure 9, Page 29. Flow Chart describing the methodology used to convert the well
logs from being sampled in depth to being sampled in time.
Figure 10, Page 31. An example of the calculation of the Porosity Logs used in the
training of the neural network and multilinear regression analysis. If the Gamma Ray (a)
v
value is below a threshold value, in this case 63 GAPI, the lithology is interpreted to be
sandstone. This is shown by shading yellow. Above 63 GAPI, the rock is interpreted to
be shale. The fluid type, either water or gas, saturating the rock is determined from the
resistivity log (b). In this well, if the resistivity is above 2 OHMM, shaded red, the fluid
is interpreted to be gas, with a density of 0.42 g/cc. If the resistivity is below the
threshold, the fluid saturant is interpreted to be water, with density of 1.1 g/cc. For each
sample, this information is combined with the density log (c) to create the porosity log
(d). We determined that the predictive power of the methods being studied is greatly
enhanced if a constant porosity value of 0.17 (e) is used wherever it a shale has been
interpreted to be the rock type. The smoothed log in each case is shown in red. The
porosity logs were calculated using the unfiltered logs, however.
Figure 11, Page 33. Architecture of the Feed Forward Network. Seven normalized
input attributes are fed into twenty-two neurons in the hidden layer. The one output
represents a porosity sample.
Figure 12, Page 34. These attributes (a-g) are combined in order to create a synthetic
porosity log that resembles as closely as possible the target log at right (h).
Figure 14, Page 37. Hinton plot that shows the strength of the weights between the input
layer and the hidden layer. The size of the box corresponds to the amplitude of the
weight, while the color indicates polarity. A red weight is negative, and a green weight is
positive.
Figure 15, Pages 39-41. Test wells from the training sequence of the Neural Network
with 1000 Epochs of Training. The CC value indicates the normalized cross-correlation
coefficient between the neural network output and the target values.
Figure 16, Pages 41-42. Performance of ANN as a function of the number of wells used
in the training set. a) Maximum Error Percentage in the test case. Note the asymptotic
appearance of the curve. Error does decrease with an increasing number of wells, but the
returns are diminishing. b) The mean testing error shows the same trend as in part a). c)
The cross correlation coefficient of the test set increases substantially with the number of
wells in the training set. As the network is exposed to a greater variety of input-output
pairs, it has more experience on which to call in order to make a predictive decision.
Figure 17, Page 43-44. Results of ANN porosity prediction from a Seismic Line passing
through well A04BP1 at Trace 2206. a) InLine showing the Structure of the Nm1. b)
InLine flattened on the Nm1 horizon. Time "0" corresponds to the peak trough amplitude
picked in the seismic volume which was interpreted to represent the N sand.
Figure 18, Pages 46-53. In a 100 ms interval around the N sand, 25 porosity values were
predicted by the ANN that correspond to the 25 seismic samples in the interval. a) The
maximum predicted porosity at each trace location is mapped and overlain onto the
vi
structure contours of the Nm1. b) Porosity map at twelve ms, or 3 seismic samples,
above the trough corresponding to the Nm1. c) Porosity map at eight ms, or 2 seismic
samples, above the trough corresponding to the Nm1. d) Porosity map at four ms, or one
seismic sample, above the trough corresponding to the Nm1. e) Porosity map at the trough
corresponding to the Nm1. f) Porosity map at four ms, or one seismic sample, below the
trough corresponding to the Nm1. g) Porosity map at eight ms, or 2 seismic samples,
below the trough corresponding to the Nm1. h) Porosity map at twelve ms, or 3 seismic
samples, below the trough corresponding to the Nm1.
Figure 19, Pages 56-58. Test wells from the training sequence of the Multivariate Linear
Regression. The CC value indicates the normalized cross-correlation coefficient between
the neural network output and the target values.
Figure 20, Page 59. Results of the Regression porosity prediction from a Seismic Line
passing through well A04BP1 at Trace 2206. a) InLine showing the Structure of the Nm1.
b) InLine flattened on the Nm1 horizon. Time "0" corresponds to the peak trough
amplitude picked in the seismic volume which was interpreted to represent the N sand.
Figure 21, Pages 60 -67. In a 100 ms interval around the N sand, 25 porosity values
were predicted by multivariate linear regression that correspond to the 25 seismic
samples in the interval. a) The maximum predicted porosity at each trace location is
mapped and overlain onto the structure contours of the Nm1. b) Porosity map at twelve
ms, or 3 seismic samples, above the trough corresponding to the Nm1. c) Porosity map at
eight ms, or 2 seismic samples, above the trough corresponding to the Nm1. d) Porosity
map at four ms, or one seismic sample, above the trough corresponding to the Nm1. e)
Porosity map at the trough corresponding to the Nm1. f) Porosity map at four ms, or one
seismic sample, below the trough corresponding to the Nm1. g) Porosity map at eight ms,
or 2 seismic samples, below the trough corresponding to the Nm1. h) Porosity map at
twelve ms, or 3 seismic samples, below the trough corresponding to the Nm1.
Figure 22, Pages 68-69. a) Histogram showing the distribution of porosity values in the
training set. Note the range of values extends to the upper limit of 0.33. b) Histograms
showing the distribution of predicted porosity values in maps shown above. The results
given by the neural network demonstrate roughly the same distribution as the porosity
training set. The results from regression analysis, however, lack the high-porosity
information.
vii
LIST OF TABLES
Table 1, Page 2. Results from previous porosity prediction studies using ANN. CC Test
is the average cross correlation coefficient between the prediction and target sets in the
test case. CC Training is the cross correlation coefficient between the prediction and
target sets in the training case. The mean testing error documents the mean error between
the prediction and target sets in the test case. Two types of ANN have been used in these
previous studies. PNN refers to a Probabilistic Neural Network, while MLFN refers to a
Multi-Layer Feed Forward Network.
Table 2, Page 3. Results from previous porosity prediction studies using a Multilinear
Regression of Seismic Attributes. CC Test is the average cross correlation coefficient
between the prediction and target sets in the test case. CC Training is the cross
correlation coefficient between the prediction and target sets in the training case. The
mean testing error documents the mean error between the prediction and target sets in the
test case.
Table 4, Page 36. Results of porosity prediction after 1000 epochs of training. All wells
except the one indicated were used in the training set. For example, the second row of
the table contains the data for the A01. All of the wells except for the A01 were used to
train the network; A01 was left out of the training as the test case. The maximum and
average errors are in units of porosity (V/V). CC Test and CC Training are the
normalized cross correlation coefficients between the prediction and the targets for the
test and training sets, respectively.
Table 5, Page 54. Results of porosity prediction from multivariate linear regression. All
wells except the one indicated were used in the training set. For example, the second row
of the table contains the data for the A01. All of the wells except for the A01 were used
to train the network; A01 was left out of the training as the test case. The maximum and
average errors are in units of porosity (V/V). CC Test and CC Training are the
normalized cross correlation coefficients between the prediction and the targets for the
test and training sets, respectively.
Table 6, Page 74. Mathematical Variables used in the paper and their abbreviations.
Table 7, Page 75. Prediction results of an ANN trained with validation training. In all
cases, it performed more poorly than a network with the same architecture but trained for
1000 epochs.
viii
ACKNOWLEDGMENTS
Ertekin for their unwavering support during this degree, as well as Professor Andy
Nyblade for agreeing to act on my committee. I would also like to thank: Eleanor for
putting up with me; Matt for helping me move; Al and Dad for rescuing me from the
winter more than once; Mom for the home cooked meals; and the dudes.
ix
INTRODUCTION
problem for geologists and geophysicists. Seismic data are often used to delineate the
structure of reservoir bodies, but are not often used to estimate the spatial distribution of
reservoir and rock properties. In this study, we compare two methods of predicting
Schultz et. al., in 1994, were the first to propose the idea of using multiple seismic
attributes to predict log properties away from well control [1]. Well-to-seismic ties are
considered to be the training points for the method; both the seismic response and the
well log property are known at the select well locations. A statistical relationship, either
linear or nonlinear, is developed at the well sites to relate the seismic response to the well
log response. We describe the use of an Artificial Neural Network (ANN) and
Geoscientists have been among the pioneers in developing uses for Artificial
Neural Networks. In many ways, neural networks are perfectly suited for geostatistical
ways and with no analytical relationships known as of yet. As computers have increased
in computational speed and power, large scale data processing with neural networks has
become feasible.
Artificial Neural Networks (ANN) have been used to predict core properties from
well logs [2], well log to well log transformations, and have used seismic properties to
predict lithology [3-6], sonic logs and shale content [7], shale stringers in a heavy oil
1
reservoir [8], density and gamma ray logs [8, 9], spontaneous potential [10, 11],
Table 1. Results of Past Studies Using ANN to Predict Porosity from Seismic
Attributes
Type of
Mean Testing
Author CC Test CC Training Network
Error (V/V)
Used
Pramanik [13] 0.86 0.95 0.037 PNN
Hampson [18] 0.62 0.95 NA PNN
Leiphardt [16] 0.62 0.82 0.026 PNN
Dorrington [19] 0.57 0.86 NA MLFN
Table 1. Results from previous porosity prediction studies using ANN. CC Test is the
average cross correlation coefficient between the prediction and target sets in the test
case. CC Training is the cross correlation coefficient between the prediction and target
sets in the training case. The mean testing error documents the mean error between the
prediction and target sets in the test case. Two types of ANN have been used in these
previous studies. PNN refers to a Probabilistic Neural Network, while MLFN refers to a
case. In these circumstances, log properties are estimated from a linearly weighted sum
of a number of seismic attributes. This was first demonstrated to yield accurate results by
2
Table 2. Results of Past Studies Using Multilinear Regression to Predict Porosity
from Seismic Attributes
Mean Testing
Author CC Test CC Training
Error (V/V)
Pramanik [13] 0.56 0.66 0.056
Hampson [18] 0.60 0.69 NA
Leiphardt [16] 0.63 0.74 0.026
between the prediction and target sets in the test case. CC Training is the cross
correlation coefficient between the prediction and target sets in the training case. The
mean testing error documents the mean error between the prediction and target sets in the
test case.
There are several advantages of these new algorithms over the conventional
inversion methods: they predict logs other than acoustic impedance; they may use seismic
attributes other than the conventional post-stack volume; they do not rely on any
particular forward model; knowledge of the seismic wavelet is not required; they may
enhance resolution; and they use testing and cross-validation as a measure of success [1,
16-18].
3
Theoretical Background
input data and target outputs. The nature of this nonlinear mapping depends upon the
type of artificial neural network used. This thesis focuses on feed forward back
layer and a variable number of hidden layers (Figure 1) [20]. These layers are each
composed of a number of neurons, which are each connected to every other neuron of the
Figure 1. Architecture of a simple Neural Network with one hidden layer. In this scenario, the
input layer is O1, the hidden layer is O2, and the output layer is O3. Each circle represents a
neuron. The input layer has four neurons, the hidden layer three, and the output one. This
4
The neurons in the input layer do not perform any computation, as they are simply
the input gate. The neurons in the hidden and output layers have weights and biases
connecting them to the neurons in the previous layer [18, 20]. Each neuron sums the
weighted and biased input from each neuron in the previous layer and then filters the sum
with a transfer function. We assume that the relationship between the inputs and target
function, the hyperbolic tangent sigmoid (Figure 2). The hyperbolic tangent sigmoid
e n − e −n
tanh(n) = n (1)
e + e −n
Figure 2. Graphic display of the hyperbolic tangent sigmoid transfer function. Note how
quickly the function saturates for values greater than five. To make full use of the shape of this
5
The output layer is the weighted and biased sum from the output of the neurons in
the last hidden layer. This sum is not filtered, so the transfer function is described as
If the output of layer m is denoted as Om, then the output of the input layer is just
O1 = I (2)
For simplicity, and since this applies directly to the network that is to be used in this
study, assume that there is only one hidden layer. The output of the jth neuron in this
= f 2 ⎜ ∑ I i wij + bij ⎞⎟
⎛ k
O2 j (3)
⎝ i =1 ⎠
Where wij and bij are the weights and biases, respectively, that connect the ith neuron in
the input layer to the jth neuron in the hidden layer [20].
The final output layer has only one neuron, corresponding to one property
prediction. If there are n neurons in the hidden layer, then the final output, which is not
n
O3 = ∑ O2 j w j + b j (4)
j =1
Where, once again, wj and bj are the weights and biases which connect the jth neuron in
6
In order to accurately map the inputs to the output, the network will continuously
update the weights and biases which connect each of the neurons until some performance
criterion is achieved. This process is called training. Many different training algorithms
exist, but error backpropagation is the most popular learning algorithm for multi-layered
An ANN minimizes the difference between the prediction and targets in any
between the network output and the targets that is used as the objective function, and is
the quantity that is to be minimized. In this case, however, it was demonstrated that
using the mean absolute error as the performance criterion yielded much more accurate
results [21]. We will now describe backpropagation using mean absolute error, and the
gradient descent method with momentum and a variable learning rate is used to train the
neural network.
For prediction-target pairs, (p1,L1), (p2,L2), … (pN, LN), we want to minimize the
1 N
E = ∑ Lj − p j (5)
N j =1
This can be accomplished by the gradient descent algorithm. To descend towards the
local minimum in a function, one can take steps in the direction of the negative gradient
For example, if γ is sufficiently small, then xn+1 will be closer to the local
minimum in F then was xn. Note that the superscripts do not represent exponents.
x n +1 = x n − γ∇F ( x n ) (6)
7
This is an iterative process, which in many circumstances will converge towards the local
minimum.
The objective, therefore, is to move towards the minimum in the mean absolute
error surface, E. Weights and biases, w and b, are iteratively updated until this minimum
is reached. This is done for all weights and biases in the network for a number of
iterations until some stopping criterion is reached. The gradient descent algorithm in this
case becomes:
∂E ( wijn )
wijn +1 = wijn − γ (7)
∂wijn
∂E (bijn )
bijn +1 = bijn − γ (8)
∂bijn
The value of γ can change between successive iterations; this is called an adaptive
learning rate. Convergence can be sped up if the learning rate is increased on flat parts of
the error surface, and decreased where the slope is steep. To implement this simply, the
learning rate is increased if the error decreases, and is decreased if the error increases.
convergence. This will essentially act as a low pass filter to smooth out any oscillations
in the convergence trajectory. To illustrate momentum learning, recall that the weight
∂E ( wijn )
Δwijn = −γ (9)
∂wijn
8
With momentum learning, this becomes:
∂E ( wijn )
Δwijn = αΔwijn −1 − (1 − α )γ (10)
∂wijn
For some α that satisfies
Weights and biases will now converge quickly and stably towards the minimum
in the mean absolute error surface [20]. In this way, neural networks can create a
transformation that minimizes the error between the output of the network and the target
mapping.
9
Multivariate Linear Regression
This method aims to combine many inputs, in this case many seismic attributes, in
a linear manner to create a mapping to the target output [16-18]. It is a simple extension
Assume that there are m inputs each of length n, with the nth sample of the mth
input denoted as Amn. These inputs are to be mapped to an output log of length n, with
M (12)
L = AW (14)
10
The weights, w, in this equation may be derived by minimizing the difference between
the prediction and the target log, in a least squares sense. The mean squared prediction
error is:
1n
E = ∑ ( Li − w0 − w1 A1i − w2 A2i − ... − wm Ami )2
2
(15)
n i=1
The minimum of this summation will occur where the first derivative with respect to the
∂E 2
=0
∂w0
M (16)
∂E 2
=0
∂wm
Finding these partial derivatives and rewriting gives:
n n n n
nw0 + ∑ w1 A1i + ∑ w2 A2i + K + ∑ wm Ami = ∑ Li
i =1 i =1 i =1 i =1
n n n n n
∑ w0 A1i + ∑ w1A1i + ∑ w2 A2i A1i +K+ ∑ wm AmiA1i = ∑ Li A1i
2
i =1 i =1 i =1 i =1 i =1
M (17)
n n n n n
∑ w0 Ami + ∑ w1 A1i Ami + ∑ w2 A2i Ami + K+ ∑ wm Ami = ∑ Li Ami
2
i =1 i =1 i =1 i =1 i =1
11
Once again, in matrix notation this becomes:
⎡ n n n n n ⎤
∑ A1i ∑ A2i ∑ A3i L ∑ Ami
⎢ ⎥w ⎡ n ⎤
⎥⎡ 0 ⎤ ⎢ i∑
i=1 i=1 i=1 i=1
⎢n Li
⎥
∑ AmiA1i ⎥⎢ w1 ⎥ ⎢ n =1 ⎥
n 2 n n n
⎢ 1i
∑ A ∑ A ∑ A A ∑ A3i A1i L
⎢ in=1 i=1
1i
i=1
2i 1i
i=1 i=1 ⎥⎢ ⎥ (18)
n n 2 n n ⎢ w ⎥ = ⎢ ∑Li A1i ⎥
⎢ ∑ A2i ∑ A1i A2i ∑ A2i ∑ A3i A2i L ∑ AmiA2i ⎥⎢ ⎥ ⎢ i=1
2
⎥
⎢i=1 i=1 i=1 i=1 i=1 ⎥ M ⎢ M ⎥
⎢ M M M M O M ⎥⎢ ⎥ ⎢∑ n
⎥
⎢ n n n n n 2 ⎥⎢ ⎣wm⎥⎦ ⎢⎣i=1Li Ami⎥⎦
⎢⎣i∑
=1
Ami ∑ A1i Ami ∑ A2i Ami ∑ A3i Ami
i=1 i=1 i=1
L ∑ Ami ⎥
i=1 ⎦
In order to solve this system for the weights, we must invert the coefficient matrix on the
−1
⎡ n n n n n ⎤
∑ A1i ∑ A2i ∑ A3i L ∑ Ami
⎢ ⎥ ⎡ nL ⎤
⎡ w0 ⎤ ⎢ n i=1 i=1 i=1 i=1
⎢ w ⎥ ⎢ ∑A
n 2 n n n ⎥ ⎢ i∑ =1
i ⎥
∑ A1i ∑ A2i A1i ∑ A3i A1i L ∑ AmiA1i ⎥ ⎢n ⎥ (19)
⎢ 1 ⎥ ⎢ i=1 1i i=1 i=1 i=1 i=1 ⎥
⎢ w2 ⎥ = ⎢ n n n 2 n n ⎢ ∑ i 1i ⎥
L A
⎢ M ⎥ ⎢ 2i ∑ A ∑ A A ∑ A ∑ A3i A2i L ∑ AmiA2i ⎥ ⎢ i=1 ⎥
i=1 i=1
1i 2i
i=1
2i
i=1 i=1 ⎥ ⎢n M ⎥
⎢ ⎥ ⎢ M M ⎥
⎢⎣wm⎥⎦ ⎢ n
M M M O ⎢ ∑Li Ami⎥
n 2 ⎥
n n n
⎣⎢i=1 ⎦⎥
⎢⎣i∑=1
Ami ∑ A1i Ami ∑ A2i Ami ∑ A3i Ami
i=1 i=1 i=1
L ∑ Ami ⎥
i=1 ⎦
The coefficient matrix is square and is invertible numerically. In this manner the weights
which minimize the least squares prediction error can be determined. While this
expensive [16-18].
Note that if
12
⎡1 A11 A21 A31 L Am1 ⎤
⎢1 A12 A22 A32 L Am 2 ⎥
⎢ ⎥
A = ⎢1 A13 A23 A33 L Am3 ⎥ (20)
⎢M
⎢ M M M O M ⎥⎥
⎢⎣1 A1n A2 n A3n L Amn ⎥⎦
Then
⎡ n n n n n ⎤
∑A1i ∑A2i ∑A3i L ∑Ami
⎢ i=1 i=1 i=1 i=1 ⎥
⎢n n 2 n n n ⎥
⎢ ∑A1i ∑A1i ∑A2i A1i ∑A3i A1i L ∑AmiA1i ⎥
⎢i=1 i=1 i=1 i=1 i=1 ⎥
AT A= ⎢ n n n 2 n n (21)
∑A2i ∑A1i A2i ∑A2i ∑A3i A2i L ∑AmiA2i ⎥
⎢i=1 i=1 i=1 i=1 i=1 ⎥
⎢ M M M M O M ⎥
⎢ n n n n n 2 ⎥
∑ A ∑ A A ∑ A A ∑ A A
⎢⎣i=1 mi i=1 1i mi i=1 2i mi i=1 3i mi L ∑ Ami ⎥
i=1 ⎦
This matrix is equal to the coefficient matrix which must be inverted in equation 19. This
matrix is also a very powerful tool used in data analysis, and is known as the covariance
matrix.
Also. note that the right hand side of Equation 19 is equal to:
⎡ ∑n L ⎤
⎢ i =1 i ⎥
⎢ n ⎥
AT L = ⎢ i∑ Li A1i ⎥ (22)
⎢ =1 ⎥
⎢n M ⎥
⎢ ∑ Li Ami ⎥
⎣⎢i =1 ⎥⎦
13
So that starting from equation 13,
AW = L (23)
this leads to
AT AW = AT L (24)
ATA is a square symmetric matrix and so in many cases may be invertible. To solve for
W = AT A( )−1 AT L (25)
These weights in vector W provide a mapping between the attributes in matrix A, and the
target log, L. The difference between the outputs of this mapping and the target log are
14
Seismic Attributes
seismic data. A seismic trace is the result of complicated interrelationships between bed
thickness, porosity, fluid saturation, lithological boundaries, and other rock properties
[15, 22, 23]. As such, a seismic trace can be transformed into an attribute which
Hundreds of seismic attributes have been developed; only some of these are well
enough understood to be quantitative, and many are redundant. We have chosen to use
Instantaneous Attributes, so called because they are calculated at every time sample of
the seismic trace. A detailed list of the attributes used, and their potential geologic
Attribute Significance
15
The instantaneous phase, frequency, and envelope of the seismic traces is
computed through complex trace analysis [26]. In complex trace analysis, a seismic
trace, s(t), is viewed as the real part of an analytical trace, S(t), that is composed of both
Where j equals the square root of negative one, s*(t) is the quadrature component of the
amplitude, known as the amplitude envelope, A(t), and a time-varying phase, θ(t), then:
S (t ) = s (t ) + js * (t ) = A(t )e iθ (t ) (26)
{ }
s (t ) = Re A(t )eiθ (t ) = A(t ) cos(θ (t )) (27)
{ }
s* (t ) = Im A(t )eiθ (t ) = A(t ) sin(θ (t )) (28)
16
A(t ) = s 2 (t ) + s*2 (t ) (29)
s * (t )
−1
θ (t ) = tan ( ) (30)
s (t )
The instantaneous frequency, f(t), is simply the time derivative of the instantaneous
1 dθ (t )
f (t ) = (31)
2π dt
17
Figure 3. a) A seismic trace in blue with its amplitude envelope in red. For every trace sample
there is an attribute sample. b) The unwrapped instantaneous phase of the trace in a). c) The
18
instantaneous frequency. d) First Derivative of the original signal with time. e) Second
Derivative of the original signal with time. f) The integrated Absolute Amplitude is also used as
19
Normalization and Principal Component Analysis
the neural network has a limited range. The domain is effectively limited as well since
the function saturates quickly for absolute input values greater than around five. To
overcome this, both the input and output data sets must be processed in order to take full
This is accomplished by processing the training set of the network to give both the
input and output sets a mean of zero and a standard deviation of one. For example the
The same transformation is applied to the output set. This means that the network
will be trained to yield outputs that have zero mean and unity standard deviation. These
normalized results are then transformed back by multiplying by the standard deviation
transform it using principal component analysis. While the input set may be large, many
has three effects: it orthogonalizes the input vectors so that they are uncorrelated, it
orders the resulting orthogonalized components such that those with the largest variation
come first, and it eliminates the components that add the least amount of variation to the
20
data set [28]. In this case, we have chosen to eliminate data that contributes less than five
normalized input set, find its unit eigenvectors, sort the eigenvectors in order from largest
to smallest eigenvalue, and then eliminate the undesired components. These remaining
eigenvectors are now used as a transformation matrix, and essentially act as the axes in a
⎡ s1 ⎤
⎡ p1 ⎤ ⎡- 0.2547 - 0.0047 0.5835 0.6075 - 0.2103 - 0.0464 0.3314 0.1652 ⎤ ⎢ ⎥
⎢ p ⎥ ⎢ 0.0298 0.0292 0.1552 0.1314 0.6962 0.6834 0.0528 - 0.0488 ⎥ ⎢ s 2 ⎥
⎢ 2⎥ ⎢ ⎥ ⎢ s3 ⎥
⎢ p 3 ⎥ ⎢ 0.8177 0.1778 0.2053 0.1183 0.2726 - 0.4002 0.0818 - 0.0493⎥ ⎢ ⎥
⎢ ⎥=⎢ ⎥ ⎢s 4 ⎥
p
⎢ ⎥ ⎢4 - 0.074 0.6729 - 0.2827 - 0.3194 0.0227 0.0362 0.5938 0.0722 ⎥ ⎢s ⎥
⎢ p 5 ⎥ ⎢ 0.0909 - 0.1316 - 0.173 0.0357 0.1057 - 0.0027 - 0.0241 0.9651 ⎥ ⎢ 5 ⎥
⎢ ⎥ ⎢ ⎥ ⎢s6 ⎥
p
⎢ ⎥ ⎢6 - 0.0686 0.6658 0.121 - 0.0085 - 0.0249 0.0374 - 0.7241 0.104 ⎥ ⎢s ⎥
⎢⎣ p 7 ⎥⎦ ⎢⎣- 0.0047 - 0.1063 0.6838 - 0.7029 - 0.0517 0.0436 0.0495 0.1416 ⎥⎦ ⎢ 7 ⎥
⎣ s8 ⎦
The above equation shows the actual transformation matrix used in determining the
principal components, pi, of the eight normalized seismic attributes, si, used in the study.
Since all of the eigenvectors are orthogonal, this is the most efficient way to
represent the data. This coordinate system now effectively communicates where along
21
CASE STUDY AND METHODOLOGY
This study makes use of data over Auger Field in the deepwater Gulf of Mexico
(GB 426, 427, 470, 471). Several studies of the geologic character of the field have been
and N Sands in Auger Field, Gulf of Mexico”, which lays the stratigraphic framework for
This study focuses on the shallowest reservoir in Auger Field, the N Interval.
Hydrocarbons in this interval are trapped by a three-way closure against the Auger Salt
Ridge and the East Auger Fault. In the majority of the reservoir, the N Interval is
composed of two massive sands, the NM1, and the NM2, termed the Lobe Facies by Bohn
(2008). Each of these is capped by a laminated member, the NL1 and the NL2,
respectively, which are termed the Levee Facies [33]. Close to the salt dome, however,
the N Interval is comprised mainly of the NM1 and NL1 members, as the NM2 and NL2
members have onlapped the Auger Salt Ridge and pinched out.
source. This incision and bypass to the neighboring Andros Basin created a valley which
thickens and widens to the south towards the depocenter. The evidence of this incision is
seen downdip and to the west, where a third facies, the Incised Fill Facies, is present
(Figure 14). The seismic signature of the interval becomes more chaotic; this represents
the character of the Incised Fill Facies, which is composed of a lower net-to-gross sand
than is present in the Lobe Facies. The NL1 and NL2 members are also seen to thicken at
the valley margin. This is most likely due to overbank deposition by the incision.
22
There is log evidence that there is overall a lower average porosity in the incised
Figure 4. Structure Map on top of the N interval, outlining the Lobe Facies, Incised Fill, and
well penetrations. Wells used in the training set are filled with red. The wells penetrating the
Incised Fill that are used in this study include the A02BP1, the A05, and 426STBP1. The A01,
23
Figure 5. a) Structure map of the NM1 with a facies interpretation overlain. The Incised Fill
Facies is shown in green; it widens and thickens to the south. b) Time slice showing lobate
geometry and the basin axis channel. c) Type well logs showing considerable change in facies
character going from the Lobe Facies (A04BP1) to the Incised Fill Facies (A02BP1).
There is extensive 3D seismic coverage over the field. The N Interval has 47 well
penetrations, but only 14 wells were drilled before production began and have a suitable
The seismic survey used was completed in 1990 before the field came on-line,
and before the Tension Leg Platform was present. The survey was shot in a North-South
orientation with three 1545 cu. in. airguns, at 2000 PSI. The far offset was 20000 feet,
24
with a CMP bin size of 100 ft. X 100 ft. giving a nominal fold of 30. The data underwent
25
Well Log Pre-Processing
In order for the well logs to be directly compared to the seismic data, a
considerable amount of pre-processing was necessary. The primary processing steps are
The despiking and quality control of the well logs was accomplished by filtering
the logs with a 35 point running mean filter. Applying the running mean filter has
striking results that are visually evident in the log character (Figure 6).
Figure 6. Comparison of the original, unedited porosity log from well A01 (in blue), and the log
The next pre-processing step is that of depth to time conversion. While the
seismic traces are sampled in units of time, the well logs are sampled in units of depth.
26
Since the goal is to create synthetic well logs based on the seismic data, it is prudent to
Generally, as is the case for the wells being studied here, the seismic transit time
to certain depths along the wellbore is measured directly in what is known as a checkshot
survey. These data provides the most accurate transformation from depth to time. The
checkshot coverage along the wellbore is sparse, and must be upsampled. The seismic
survey being used has a sampling interval of 4 ms. For twenty samples around the Nm1,
giving a 100 ms window, depths are interpolated to every 4 ms sample through spline
Now that the depth at which each seismic sample occurs is estimated, we must
now be certain that a well log sample also occurs at that depth. For the most part, this is
not the case, and we must interpolate the well log to give us a sample at that depth
(Figure 8). The extent of these points on the well log corresponds to a 100 ms seismic
27
Figure 7. Example of the interpolation used to calculate the depth at which each seismic sample
of interest occurs. This is data from the checkshot from well A01. At this well, the Nm1 occurs at
4.580 seconds two-way-time. The depths are calculated for every 4 ms in a 100 ms window
28
Figure 8. Porosity values that have been calculated via a spline interpolation for each of the
Figure 9. Flow Chart describing the methodology used to convert the well logs from being
29
Porosity Calculation from Density Logs
Although neutron porosity logs were available for analysis, it was uncertain
whether each logging tool had been calibrated to the same fluid and rock type, and no
information was available that described the overall calibration technique. Besides, it is
generally accepted that porosity calculation from bulk density logs is more accurate[34].
To calculate porosity, φ, we use the rock matrix density, ρm, the fluid density, ρf,
and the bulk density well log, ρb. The average rock density in the sandstones from core
reports from four exploration wells is 2.66 g/cm3. The average rock density in the shales
hydrocarbons. This was determined by the electrical resistivity log. If the resistivity was
above an interpreted threshold value, than the hydrocarbon value was used; if below the
threshold, the water value was used. The hydrocarbon density calculated from
composition and phase considerations is 0.42 g/cc at 10,500 PSI, which is roughly the in
situ reservoir pressure. The water density used was 1.1 g/cc. The calculation is then:
ρ m − ρb
φ= (23)
ρm − ρ f
30
Figure 10. An example of the calculation of the Porosity Logs used in the training of the neural
network and multilinear regression analysis. If the Gamma Ray (a) value is below a threshold
value, in this case 63 GAPI, the lithology is interpreted to be sandstone. This is shown by shading
yellow. Above 63 GAPI, the rock is interpreted to be shale. The fluid type, either water or gas,
saturating the rock is determined from the resistivity log (b). In this well, if the resistivity is
above 2 OHMM, shaded red, the fluid is interpreted to be gas, with a density of 0.42 g/cc. If the
resistivity is below the threshold, the fluid saturant is interpreted to be water, with density of 1.1
g/cc. For each sample, this information is combined with the density log (c) to create the
porosity log (d). We determined that the predictive power of the methods being studied is
greatly enhanced if a constant porosity value of 0.17 (e) is used wherever it a shale has been
interpreted to be the rock type. The smoothed log in each case is shown in red. The porosity logs
31
Porosity Prediction with an Artificial Neural Network
Once the well logs and seismic attributes had been processed as described in
Chapter 2, an ANN could be suitably trained to generate a suite synthetic well logs from
appropriate architecture was found, dozens of networks were designed, tested, and
ultimately rejected. These networks varied in the number of hidden layers, the number of
neurons in each of those hidden layers, training algorithms, transfer functions, and
optimization criteria. Among the many networks tested, the most accurate and well-
consisting of one hidden layer of 22 neurons with tangent sigmoid transfer functions that
used a gradient-descent with momentum and adaptive learning rate training algorithm.
The performance criterion was set to minimize the mean absolute error, in contrast to the
more standard mean square error. This choice was once again heuristic; the network was
32
Figure 11. Architecture of the Feed Forward Network. Seven normalized input attributes are
fed into twenty-two neurons in the hidden layer. The one output represents a porosity sample.
All of the logs and the corresponding seismic data were compiled into one set.
The seismic attributes were normalized and underwent principal component analysis
which reduced the number of inputs from eight to seven. The porosity targets were also
normalized to a mean of zero. The training set consisted of the well log and seismic data
from all but one well, which was separated as the test set.
33
Figure 12. These attributes (a-g), also listed on page 18, are combined in order to create a
synthetic porosity log that resembles as closely as possible the target log at right (h).
We rotated through the wells, successively taking one aside for the test set and
training a new network with the same network architecture in order to appropriately
gauge the predictive power of this network architecture. Training commenced with both
validation stopping and with stopping after 1000 epochs. When validation stopping was
used, every sixth sample of the training set was set aside for validation. Results for each
training cycle are shown in Table 3, where the well indicated is the well used in the test
set.
34
Figure 13. Flow Chart Describing Training Method.
35
RESULTS
Table 4. Results of porosity prediction after 1000 epochs of training. All wells except the one
indicated were used in the training set. For example, the second row of the table contains the
data for the A01. All of the wells except for the A01 were used to train the network; A01 was left
out of the training as the test case. The maximum and average errors are in units of porosity
(V/V). CC Test and CC Training are the normalized cross correlation coefficients between the
prediction and the targets for the test and training sets, respectively.
The performance of the ANN was analyzed by keeping one well aside as a test
case, and training the network on the remaining thirteen wells. This procedure was
performed fourteen times, each time testing the network on a new well.
The testing results demonstrate that the network generalizes well, and has
impressive predictive powers. With an average cross correlation coefficient of 0.72 for
36
the test case, this network outperforms many of those from previous studies (see Intro)
[16, 18, 19]. The mean error is on the order of 3% porosity which is much lower than
This was accomplished without the use of an inverted acoustic impedance volume
as input, which has been demonstrated to have the highest correlation with porosity of
any single attribute [19]. We attribute this success to the quality control of the well logs
and well-ties, the preprocessing of the input seismic attributes and output porosity logs
which includes normalization and principal component analysis, and the choice of
training algorithm.
The Hinton Plot below shows that the normalized inputs are given roughly equal
weighting when estimating the porosity log. This suggests that all of the attributes used
in the input set have a functional relationship with porosity, and that that relationship may
be nonlinear in nature.
37
Figure 14. Hinton plot that shows the strength of the weights between the input layer and the
hidden layer. The area of the box corresponds to the amplitude of the weight, while the color
indicates polarity. A red weight is negative, and a green weight is positive. The seven inputs
being weighted are the principal components of the original, normalized seismic
attributes.
38
39
40
Figure 15. Test wells from the training sequence of the Neural Network with 1000 Epochs of
Training. The CC value indicates the normalized cross-correlation coefficient between the neural
Instead of removing only one well as the test case, we successively removed additional
wells until only one well served as the training case. As expected, the predictive power
of the ANN grew as the number of training wells increased. Both absolute error, as well
as error percent decreased, while the testing cross correlation coefficient increased. In
41
Figure 16. Performance of ANN as a function of the number of wells used in the training set. a)
Maximum error percentage in the test case. Note the asymptotic appearance of the curve. Error
42
does decrease with an increasing number of wells, but the returns are diminishing. b) The mean
testing error shows the same trend as in part a). c) The cross correlation coefficient of the test set
increases substantially with the number of wells in the training set. As the network is exposed to
a greater variety of input-output pairs, it has more experience on which to call in order to make a
predictive decision.
43
Figure 17. Results of ANN porosity prediction from a Seismic Line passing through well
A04BP1 at Trace 2206. a) InLine showing the Structure of the Nm1. b) InLine flattened on the Nm1
horizon. Time "0" corresponds to the peak trough amplitude picked in the seismic volume which
Mapped Results
Porosity was predicted in a 100 ms interval surrounding the Nm1. Since the
sampling rate is 4 ms, there are 25 porosity samples generated for each trace. One can
then construct 25 maps connecting corresponding porosity samples. Those results are
shown below for predictions from both the ANN and the multilinear regression. The
network and weights that were used came from the training sequence that used the
constant value of porosity of 0.17 for shales, since this method was deemed to be more
44
Note that labels of negative times show a map above the interval of maximum
seismic amplitude corresponding to the Nm1, while positive numbers refer to maps
showing porosity values below the time of maximum amplitude in the seismic trace.
Both methods show clear indications of fan lobes. (See the two figures
immediately below) The results may even suggest a fining upward facies, as the coarsest
and most porous part of the sand package occurs at the bottom of the Nm1 (+8-12 ms).
This fining upward characteristic is noticed in the well logs, and is indicative of turbiditic
origins.
The maps generally show an area of high porosity to the southwest of the salt
diapir. This corresponds to the thickness area of the sand, as indicated by both an
isochron map and well penetrations. Directly to the northeast of that, the sand thins and
onlaps the salt, and the porosity decreases. We believe that these results are geologically
Figure 18. In a 100 ms interval around the N sand, 25 porosity values were predicted by
multivariate linear regression that correspond to the 25 seismic samples in the interval. a) The
maximum predicted porosity at each trace location is mapped and overlain onto the structure
contours of the Nm1. b) Porosity map at twelve ms, or 3 seismic samples, above the trough
corresponding to the Nm1. c) Porosity map at eight ms, or 2 seismic samples, above the trough
corresponding to the Nm1. d) Porosity map at four ms, or one seismic sample, above the trough
corresponding to the Nm1. e) Porosity map at the trough corresponding to the Nm1. f) Porosity
map at four ms, or one seismic sample, below the trough corresponding to the Nm1. g) Porosity
map at eight ms, or 2 seismic samples, below the trough corresponding to the Nm1. h) Porosity
map at twelve ms, or 3 seismic samples, below the trough corresponding to the Nm1.
45
18a
46
18b
47
18c
48
18d
49
18e
50
18f
51
18g
52
18h
53
Multivariate Linear Regression Results with Constant Shale Porosity
Table 5. Results of porosity prediction from multivariate linear regression. All wells except the
one indicated were used in the training set. For example, the second row of the table contains the
data for the A01. All of the wells except for the A01 were used to train the network; A01 was left
out of the training as the test case. The maximum and average errors are in units of porosity
(V/V). CC Test and CC Training are the normalized cross correlation coefficients between the
prediction and the targets for the test and training sets, respectively.
The results of this analysis are in inline with past studies [16-18]. The test sets
had on average a cross correlation with the target sets of around 0.62. This is
considerably lower than the results from the ANN. Although the number is low, there is
still good visual agreement between the estimates and targets. The average error is also
54
The weights derived through analysis are:
⎡ w0 ⎤ ⎡ 0 ⎤
⎢ w ⎥ ⎢ 0.156 ⎥
⎢ 1⎥ ⎢ ⎥
⎢ w2 ⎥ ⎢ - 0.183⎥
⎢w ⎥ ⎢ ⎥
⎢ ⎥=⎢
3 0.426 ⎥
⎢ w4 ⎥ ⎢ 0.123 ⎥
⎢ ⎥ ⎢ ⎥
w
⎢ ⎥ ⎢
5 0.133 ⎥
⎢ w6 ⎥ ⎢0.0142⎥
⎢ ⎥ ⎢ ⎥
⎣ w7 ⎦ ⎣- 0.172⎦
Keep in mind that these weights operate on normalized and transformed inputs.
With these weights having been derived, they were used to transform the
entire seismic volume into a porosity volume 100 ms, or 25 samples, in length around the
Nm1 horizon.
The first map shows the maximum porosity value at each trace position.
Although it predicts lower porosity values than the ANN, both methods show the same
trends. Namely, that porosity is highest in a northwest-southeast trending belt just to the
55
56
57
Figure 19. Test wells from the training sequence of the Multivariate Linear Regression. The CC
value indicates the normalized cross-correlation coefficient between the neural network output
58
Figure 20. Results of the Regression porosity prediction from a Seismic Line passing through
well A04BP1 at Trace 2206. a) InLine showing the Structure of the Nm1.
Figure 21, Pages 60 -67. In a 100 ms interval around the N sand, 25 porosity values were
predicted by multivariate linear regression that correspond to the 25 seismic samples in the
interval. a) The maximum predicted porosity at each trace location is mapped and overlain onto
the structure contours of the Nm1. b) Porosity map at twelve ms, or 3 seismic samples, above the
trough corresponding to the Nm1. c) Porosity map at eight ms, or 2 seismic samples, above the
trough corresponding to the Nm1. d) Porosity map at four ms, or one seismic sample, above the
trough corresponding to the Nm1. e) Porosity map at the trough corresponding to the Nm1. f)
Porosity map at four ms, or one seismic sample, below the trough corresponding to the Nm1. g)
Porosity map at eight ms, or 2 seismic samples, below the trough corresponding to the Nm1. h)
Porosity map at twelve ms, or 3 seismic samples, below the trough corresponding to the Nm1.
59
60
61
62
63
64
65
66
67
CONCLUSIONS
characterization by estimating the rock property away from well control. This improved
image of the reservoir allows for more advanced and accurate reservoir models,
production wells.
We have investigated two methods for the prediction of porosity from multiple
Regression. The valor of these two methods has been demonstrated on a data set from an
Overall, the ANN predicted a greater range of porosity values than did multilinear
regression (Figure 22). The ANN also outperformed the regression analysis
considerably, both in average error, but also in the test cross correlation coefficient
values. Although more computationally expensive, the ANN prediction was on average
10.4% more similar to the target than was the regression prediction.
68
Figure 22. a) Histogram showing the distribution of porosity values in the training set. Note
the range of values extends to the upper limit of 0.33. b) Histograms showing the distribution of
predicted porosity values in maps shown above, which are correspond to Figures 18e and 22e.
The results given by the neural network demonstrate roughly the same distribution as the
porosity training set. The results from regression analysis, however, lack the high-porosity
information.
The several advantages of these new algorithms over the conventional inversion
methods are that: they predict logs other than acoustic impedance; they may use seismic
attributes other than the conventional post-stack volume; they do not rely on any
69
particular forward model; knowledge of the seismic wavelet is not required; they may
enhance resolution; and they use testing and cross-validation as a measure of success [1,
16-18].
Despite these advantages, these techniques, like any statistical rock property
estimation, may not always work well. Like any data driven techniques, the quality,
uniformity, and processing of the data set will determine the effectiveness of both ANN
Future Directions
While the accuracy of both the neural network and the regression analysis has
been demonstrated through error testing and correlation studies, it is important to exercise
introduction, other methods of predicting reservoir porosity exist that are independent of
seismic data. Well testing and reservoir simulation may serve as verification for a
Spatial interpolation of the well log porosity can be very accurate in close
proximity to the well data. Further from well control, however, the accuracy can quickly
based prediction techniques could be married with seismic based prediction techniques.
Spatial data were not used in this study because of the strong correlation between lateral
position and porosity. When included in the ANN analysis, spatial data were over
emphasized by the network, ignoring needed information in the seismic data. Since this
70
study was designed to examine the possibility of using seismic data in porosity
The number of seismic attributes used in porosity prediction need not be limited
to the eight used here. Hundreds of seismic attributes are known, and new ones could be
invented solely for rock property prediction. A neural network may very well prove to be
the quickest way to determine which of the hundreds of seismic attributes correlate most
As shown in Figure 22, the two types of analysis yielded different porosity
“spectra”. The two methods attempt to minimize two different types of statistical
dispersion, namely mean squared error in the regression case, and mean absolute error in
the network case. This may account for some of the dissimilarity. It should be possible
to find appropriate linear weights which minimize mean absolute error, and therefore
may provide for a more direct comparison between the results given by the neural
network and the regression. This however, may not account for the difference in the
shape of the porosity spectra. When a completely random array of data is fed into the
neural network, for example, the output is strongly centered around the peak of 25% (see
Figure 22), but there is virtually no spread in the results whatsoever. This is an intriguing
71
Bibliography
72
20. Martin T. Hagan, H.B.D., Mark Eagle, Neural Network Design. First ed. 1996,
Boston, MA: PWS Publishing Company.
21. Taji, K., Miyake, T., Tamura, H., On Error Backpropagation Algorithm Using
Absolute Error Function. 1999 IEEE International Conference on Systems, Man,
and Cybernetics, 1999. IEEE SMC '99 Conference Proceedings. , 1999. 5: p. 401-
406.
22. Satinder, C. and J.M. Kurt, Emerging and future trends in seismic attributes.
2008, SEG. p. 298-318.
23. Satinder, C. and J.M. Kurt, Introduction to this special section---Seismic
Attributes. 2008, SEG. p. 296-297.
24. Taner, M.T., Seismic Attributes. CSEG Recorder, 2001. 26(7): p. 48-56.
25. Bruce, H., Stratigraphically significant attributes. 2008, SEG. p. 320-324.
26. Taner, M.T., F. Koehler, and R.E. Sheriff, Complex seismic trace analysis. 1979,
SEG. p. 1041-1063.
27. Ecoublet, P., and Symes, W., Application of Probabilistic Neural Network to
Porosity Prediction, in The Rice Inversion Project, Annual Report. 1998, Rice
University: Houston, TX.
28. Dunteman, G.H., Principal Components Analysis. Quantitative Applications in
the Social Sciences. 1989, Thousand Oaks, CA: Sage Publications Inc. 96.
29. Booth, J.R., et al., Paleo-bathymetric controls on the stratigraphic architecture
and reservoir development of confined fans in the Auger Basin: central Gulf of
Mexico slope. Marine and Petroleum Geology, 2003. 20(6-8): p. 563-586.
30. Booth, J.R., DuVernay III, A.E., Pfieffer, D.S., Styzen, M.J., Sequence
Stratigraphic Framework, Depositional Models, and Stacking Patterns of Ponded
and Slope Fan Systems in the Auger Basin: Central Gulf of Mexico Slope.
GCSSEPM Foundation 20th Annual Research Conference - Deep Water
Reservoirs of the World, 2000: p. 82-103.
31. McGee. D.T., B., P.W., Gary, P.S., Pfeiffer, D.S., Sheiman, J.L., Geologic Models
and Reservoir Geometries of Auger Field, Deepwater Gulf of Mexico. GCSSEPM
Foundation 15th Annual Research Conference - Submarine Fans and Turbidite
Systems, 1994: p. 245-256.
32. Kendrick, J.W., Turbidite Reservoir Architecture in the Northern Gulf of Mexico
Deepwater: Insights from the Development of Auger, Tahoe, and Ram/Powell
Fields. GCSSEPM Foundation 20th Annual Research Conference - Deep Water
Reservoirs of the World, 2000: p. 450-468.
33. Bohn, C., Reilly, M., Seren, D., and Valenti, J., Accommodation history, reservoir
architecture, and production behavior in N and O sands of Auger Field,
Deepwater Gulf of Mexico. . AAPG Bulletin, 2008.
34. Issler, D.R., A new approach to shale compaction and stratigraphic restoration,
Beaufort-Mackenzie Basin and Mackenzie Corridor, northern Canada. 1992. p.
1170-1189.
73
Table 6. Variable Definitions
Variable Description
tanh Hyperbolic Tangent Sigmoid Transfer
Function
Ii ith Sample of the input vector, I
O1 Output of the Input Layer of ANN
O2j Output of the jth neuron in the 2nd, or
hidden, layer of the ANN
f2 Transfer function associated with the 2nd,
or hidden, layer of ANN
wij Weight connecting the ith neuron in the
input layer to the jth neuron in the hidden
layer
bij Bias connecting the ith neuron in the input
layer to the jth neuron in the hidden layer
E Mean Absolute Error between target and
prediction
pj jth sample of the predicted log
Lj jth sample of target Log
γ Adaptive Learning Rate constant
α Constant used in Momentum Learning
wi Weight multiplied by the ith attribute input
in regression analysis
W Vector composed of weights wi
A Matrix of Input Attributes
E2 Mean Squared Prediction Error
AT Transpose of matrix A
S(t) Analytical Trace
s(t) Seismic Trace, the real part of S(t)
s*(t) Quadrature component of the Analytical
Trace
A(t) Amplitude Envelope of Analytical Trace
θ(t) Instantaneous Phase
f(t) Instantaneous Frequency
zi Depth to log sample i
vi Wave Velocity in interval i
ti Cumulative Transit Time to depth, zi
φ Porosity
ρm Matrix Density
ρf Fluid Density
ρb Bulk Density
74
APPENDIX
In order to demonstrate that using a fixed number of training epochs was justified,
included below are the testing results from a neural network with the same architecture
used in Chapter 3, but trained using validation training.
In all measures of success, the validation trained network performed more poorly
than the network that was simply trained for 1000 epochs. The results below, in Table 7,
can be compared to the results in Table 5.
Max Error
Well Name Max Error Mean Error Percent CC Test CC Training
A02BP1 0.119183655 45.45776074 0.035007468 0.489084205 0.571225
A04BP1 0.126220945 43.74880299 0.047002472 0.263560985 0.638067575
A05 0.145012221 85.30130652 0.03960731 0.457423974 0.669981543
A06 0.106158558 60.23677415 0.037137763 0.631343879 0.587573059
A07 0.076808279 45.176815 0.030712686 0.876160938 0.576227064
A08 0.117827823 69.31048433 0.034462862 0.763788719 0.614099391
A09 0.115866659 68.15685804 0.030843336 0.60888349 0.628635092
A13 0.101556141 59.7388582 0.028509361 0.60487601 0.612899496
A14 0.100966756 59.39220951 0.032074924 0.437389553 0.592034589
A19 0.137616819 80.95106995 0.038444913 0.344678982 0.573031905
426STBP1 0.126965017 74.68530394 0.04363576 0.8539127 0.49864843
427 0.092790164 54.58244966 0.030825303 0.462827272 0.617828737
470 0.130205815 76.59165575 0.03879009 0.654684162 0.616431442
AVERAGES 0.115167604 63.33310375 0.03592725 0.502597455 0.599744871
Table 7. Prediction results of an ANN trained with validation training. In all cases, it performed more
poorly than a network with the same architecture but trained for 1000 epochs.
75