You are on page 1of 21

ATTRIBUTES REVISITED

By:
M. Turhan Taner

Rock Solid Images


Houston, Texas
1992
(Revised Sep. 2000)
More information about Seismic Attributes can be found @ www.rocksolidimages.com such as technical papers, ATTRIB3D® product information and seismic
attribute generation and classification services

Contents
Preface
Oxford Dictionary Definition of "Attribute"

Introduction

Classification of Attributes
Pre-Stack Attributes
Post-Stack Attributes
Instantaneous Attributes
Wavelet Attributes
Physical Attributes
Geometrical Attributes
Reflective Attributes
Transmissive Attributes

Classification and Calibration Methods


Knowledge based Expert systems
Statistics of Attributes, Geostatistics
Linear Discrimination and PCA
Unsupervised Classification and Calibration
Supervised Training of, and Classification by, Neural Networks

Computational Procedures
Frequency Domain Computation
Discrete Time Domain Computation
Gabor-Morlet Decomposition
Simple Harmonic Motion Method

Computation of Basic Attributes


Envelope
Rate of Change of Envelope
Second Derivative of Envelope
Instantaneous Phase
Intensity Varying Phase Display
Normalized Amplitude Trace
Instantaneous Frequency
Weighted Mean Frequency
Thin Bed Indicator
Instantaneous Acceleration
Instantaneous Band Width
Instantaneous Dominant Frequency
Instantaneous Q Quality Factor
Relative Acoustic Impedance

Wavelet Attributes

Geometrical Attributes
Computation of Geometrical Attributes
Event Continuity
Instantaneous (Phase) Dip
Instantaneous Lateral Continuity
Attributes Computed by Dip Scanning
Maximum Lateral Semblance and corresponding Dips as Lateral Continuity Attributes
Dip of Maximum Lateral Semblance
Contents
Preface
Oxford Dictionary Definition of "Attribute"

Introduction

Classification of Attributes
Pre-Stack Attributes
Post-Stack Attributes
Instantaneous Attributes
Wavelet Attributes
Physical Attributes
Geometrical Attributes
Reflective Attributes
Transmissive Attributes

Classification and Calibration Methods


Knowledge based Expert systems
Statistics of Attributes, Geostatistics
Linear Discrimination and PCA
Unsupervised Classification and Calibration
Supervised Training of, and Classification by, Neural Networks

Computational Procedures
Frequency Domain Computation
Discrete Time Domain Computation
Gabor-Morlet Decomposition
Simple Harmonic Motion Method

Computation of Basic Attributes


Envelope
Rate of Change of Envelope
Second Derivative of Envelope
Instantaneous Phase
Intensity Varying Phase Display
Normalized Amplitude Trace
Instantaneous Frequency
Weighted Mean Frequency
Thin Bed Indicator
Instantaneous Acceleration
Instantaneous Band Width
Instantaneous Dominant Frequency
Instantaneous Q Quality Factor
Relative Acoustic Impedance

Wavelet Attributes

Geometrical Attributes
Computation of Geometrical Attributes
Event Continuity
Instantaneous (Phase) Dip
Instantaneous Lateral Continuity
Attributes Computed by Dip Scanning
Maximum Lateral Semblance and corresponding Dips as Lateral Continuity Attributes
Dip of Maximum Lateral Semblance
Smoothed dips of maximum lateral semblance
Dip Variance (Local Dip versus average Dip)
Lateral Continuity
Smoothed Maximum Lateral Semblance
Variance of maximum coherency, instantaneous versus average

Hybrid Attributes
Parallel Bedding Indicator
Chaotic zone indicator
Zones of Unconformities
Shale Indicator

FILTER DESIGN
Band Pass Filters
Butterworth Filter
Low Pass Butterworth Filter
High Pass Butterworth Filter
Band Pass Butterworth Filter
Minimum Phase Butterworth Filters
Taner Filter
Low Pass Filter
High Pass Filter
Band Pass Filter
CONVOLUTIONAL HILBERT/BUTTERWORTH FILTER
CONVOLUTIONAL DIFFERENTIATION FILTER

Preface:
This is the third edition of the "Attributes Revisited" report. We have added a new classification of the attributes, as we understand them at the present time. This classification may
well change as our understanding of the use of attributes improves.  We give a full description of each attribute and include their projected use in interpretation. The whole report is
put in an html format to be viewed interactively by the user with any Web browser. There will also be a printed copy available. This report contains all 2 -D and 3-D post-stack and 2-D
pre-stack attributes.   

Oxford Dictionary Definition of "Attribute":


A quality ascribed to any person or thing.

Definition:
Seismic Attributes are all the information obtained from seismic data, either by direct measurements or by logical or experience based reasoning.

Introduction
Based on their definition, the computation and the use of attributes go back to the origins of seismic exploration methods. The arrival times and dips of seismic events were used in
geological structure estimation. Frank Rieber in the 1940's introduced the Sonograms and directional reception. This method was extensively used in noise reduction and time
migration. The introduction of auto-correlograms and auto-convolograms (Anstey and Newman) led to better estimates of multiple generation and more accurate use of the later
developed deconvolution.  NMO velocity analysis gave better interval velocity estimates and more accurate subsurface geometries. Bright spot techniques led to gas discoveries, as
well as to some failures. This was improved by the introduction of AVO technology. Each of these developments has helped our understanding of the subsurface and reduced the
uncertainties. Unfortunately, one of the principal failures of any of the individual techniques was our implicit dependence on it. Finally, the power of the combined use of a number of
attributes is being recognized and successful techniques are being introduced. The attribute discussed in this paper is the outcome of the work relating to the combined use of
several attributes for lithology prediction and reservoir characterization..

Complex seismic trace attributes were introduced around 1970 as useful displays to help interpret the seismic data in a qualitative way. Walsh of Marathon published the first article
in the 1971 issue of Geophysics under the title of " Color Sonograms". At the same time Nigel Anstey of Seiscom -Delta had published “Seiscom 1971” and introduced reflection
strength and mean frequency. He also showed color overlays of interval velocity estimates for lithological differentiation. The new attributes were computed in the manner of radio
wave reception. The reflection strength was the result of a low pass filtered, rectified seismic trace. The color overlays showed more information than was visible on the black and
white seismic sections. Realizing the potential for extracting useful instantaneous information, Taner, Koehler and Anstey turned their attention to wave propagation and simple
harmonic motion. This led to the recognition of the recorded signal as representing the kinetic portion of the energy flux. Based on this model, Koehler developed a method to
compute the potential component from its kinetic part. Dr. Neidell suggested the use of the Hilbert transform. Koehler proceeded with the development of the frequency and time
domain Hilbert transform programs, which made possible practical and economical computation of all of the complex trace attributes. In the mid 70's three principal attributes were
pretty well established. Over the years a number of others were added.

The study and interpretation of seismic attributes give some qualitative information of the geometry and the physical parameters of the subsurface. It has been noted that the
amplitude content of the seismic data is the principal factor for the determination of physical parameters, such as the acoustic impedance, reflection coefficients, velocities,
absorption etc. The phase component is the principal factor in determining the shapes of the reflectors, their geometrical configurations etc. Our objective is to bring the interpretation
of attributes from a qualitative manner to a more quantitative manner. In this paper we will first discuss the several computational methods of conventional attributes, basically the
computation of the analytic trace.  In the second part we will present computation of the conventional attributes and their derivatives. One point that must be brought out is that we 
define all seismically driven parameters as the Seismic Attributes. They can be velocity, amplitude, frequency, rate of change of any of these with respect to time or space and so
on. We will classify the attributes based on their computational characteristics. They can be computed from pre-stack or post stack data sets. Some of the attributes computed from
the complex trace such as envelope, phase etc. correspond to the various measurements of the propagating wave front. We will call these the 'Physical Attributes'. Others,
computed from the reflection configuration and continuity, we will call 'Geometrical Attributes'. The principal objectives of the attributes are to provide accurate and detailed
information to the interpreter on structural, stratigraphic and lithological parameters of the seismic prospect.

Classification of Attributes
This paper is written to provide the background information to the RSI_ATTRIB3D interactive seismic attribute computation program developed by Seismic Research Corporation,
which later became a part of Rock Solid Images. The project was sponsored by a number of oil companies, initially by the Italian Oil Company ENI-AGIP. The initial objective was
to develop as many physical attributes as possible in order to define the lithological parameters and reservoir characteristics from different points of view. In the development we
established a general classification of attributes based on their input data and their usage. Attributes can be computed from pre -stack or from post-stack data before or after time
migration. The procedure is the same in all of these cases. Attributes can be classified in many different ways. Several authors have given their own classification (please see
references). Here we give a classification based on the characteristics of the attributes.

Pre-Stack Attributes
Input data are CDP or image gather traces. They will have directional (azimuth) and offset related information. Computations generate huge amounts of data; hence they are not
practical for initial studies.

Post-Stack Attributes
Stacking is an averaging process, losing offset and azimuth related information. Input data could be CDP stacked or migrated. One should note that time migrated data will maintain
their time relations, hence temporal variables, such as frequency, will retain their physical dimensions. For depth migrated sections, frequency is replaced by wave number, which is
a function of propagation velocity and frequency. Post-stack attributes are better for observing large amounts of data in initial investigations. For detailed studies, pre-stack attributes
may be incorporated.

Computationally, we divide attributes into two general classes.

Class I attributes are computed directly from traces. This data could be pre - or post -stack, 2 -D or 3 -D, before or after time migration. Trace envelope and its derivatives,
instantaneous phase and its derivatives, bandwidth, Q, dips etc. are some of the attributes computed this way.

Class II attributes are computed from the traces with improved S/N ratios after lateral scanning and semblance-weighted summation. Details of the computation are given in the
Maximum Semblance Computation section of the Geometrical attributes. All of the Class I attributes are computed in Class II. In addition lateral continuity and dips of maximum
semblance are computed from the scanning procedure.

Based on the information content, attributes are divided into two general categories:

Instantaneous Attributes
Instantaneous attributes computed sample by sample, representing instantaneous variations of various parameters. Instantaneous values of attributes such as trace envelope, its
derivatives, frequency and phase may be determined from complex traces. Both Class I and Class II attributes are computed.

Wavelet Attributes
Instantaneous attributes computed at the peak of the trace envelope have a direct relation to the Fourier transform of the wavelet in the vicinity of the envelope peak.  For example,
Instantaneous frequency at the peak of the envelope is equal to the mean frequency of the wavelet amplitude spectrum. Instantaneous phase corresponds to the intercept phase of
the wavelet. This attribute is also called the "response attribute". Both Class I and Class II attributes are computed.

Another classification is based on the relation of the attributes to the geology:

Physical Attributes
Physical attributes relate to physical qualities and quantities. The magnitude of the trace envelope is proportional to the acoustic impedance contrast, frequencies relate to the bed
thickness, wave scattering and absorption. Instantaneous and average velocities directly relate to rock properties. Consequently, these attributes are mostly used for lithological
classification and reservoir characterization.

Geometrical Attributes
Geometrical attributes describe the spatial and temporal relationship of all other attributes. Lateral continuity measured by semblance is a good indicator of bedding similarity as well
as discontinuity. Bedding dips and curvatures give depositional information. Geometrical attributes were initially thought to help the stratigraphic interpretation. However, further
experience has shown that the geometrical attributes defining the event characteristics and their spatial relations, quantify features that directly help in the recognition of depositional
patterns, and related lithology.

Most of the attributes, instantaneous or wavelet, are assumed to study the reflected seismic wavelet characteristics. That is, we are considering the interfaces between two beds.
However, velocity and absorption are measured as quantities occurring between two interfaces, or within a bed. Therefore, we can divide the attributes into two basic categories
based on their origin.

Reflective Attributes
Attributes corresponding to the characteristics of interfaces. All instantaneous and wavelet attributes can be included under this category. Pre-stack attributes such as AVO are also
reflective attributes, since AVO studies the angle dependent reflection response of an interface.

Transmissive Attributes
Transmissive attributes relate to the characteristics of a bed between two interfaces. Interval, RMS and average velocities, Q, absorption and dispersion come under this category.

We will define all of the available attributes in the following sections and indicate their categories and their possible relation to lithology, reservoir characteristics and depositional
settings. In most instances individual attributes may indicate several possible conditions, hence their logically combined use to minimize the inherent uncertainty. We call individual
attributes measuring only one quantity "Primitive" attributes. These primitive attributes may be logically combined to form "Hybrid" attributes. This combination is knowledge
based. We have several attributes of this form, which are described later.

Gabor-Morlet type joint Time-Frequency analysis allows us to study frequency-varying attributes. Instantaneous spectra, spectral ratio and phase differences provide measurements
for bed thickness variation, absorption and dispersion estimates.

Classification and Calibration Methods


Attribute classification helps define the combinations to be used for optimum lithology and reservoir classification. As we have seen above, there are many different ways of
computation and many different classes of attributes available. Their utilization in quantitative interpretation is the main proof of their significance. At this time, we have defined four
different methods of classification and calibration. Here we give a short description of the methods involved. Readers requiring further details should contact the author.

Knowledge based Expert systems


This method uses knowledge-based combinations and calibrations of groups of attributes with fuzzy logic to reflect the interpreter’s experience.  This type of classification can be
used for large data sets for a quick “look-see” type of interpretation, or when looking for a specific condition.

 Statistics of Attributes, Geostatistics


These represent older and well -established methods. Cross platting with linear and various non -linear scales, measuring various statistics have been used as viable tools.
Interpolation and extrapolation between and beyond wells have been improved by the introduction of Kriging. Incorporating the seismic and other soft information let to the
development of Co-kriging.  However these methods strongly depend on the estimate of probability of each factor and provide estimates of most likely situations of many kinds, from
which the interpreter has to make the final decision.  

 Linear Discrimination and PCA


Principal Component Analysis (PCA) shows the principal projections where the data has the largest variance, hence the best possible discrimination. Newly developed Independent
Component Analysis (ICA) determines the projections that are most discriminating. Linear discrimination works satisfactorily when two classes are involved and the classification
boundary is not very complicated. However, PCA and ICA are very useful analytical tools to determine the most important attribute components to be used in the non -linear
discrimination using Neural Networks.

Unsupervised Classification and Calibration


This type of approach seeks some structure in the data set. Kohonen's Self Organizing Map (SOM) method is one the most profound methods in the Artificial Intelligence and Neural
Network field. A data set may be defined by any combination of attributes and SOM generates topologically related clusters. If the selected attributes are geometrical, then the
clusters are based on the geometrical variation. The method generates coordinates of cluster centers with given attribute coordinates. However, it does not relate the cluster to any
physical or reservoir condition. This has to be done in the calibration stage.

Supervised Training of, and Classification by, Neural Networks


There have been a number of Neural Networks developed within the last couple of decades. Supervised trainable networks are used in many different fields. In this case, the user
provides some examples for the neural network to learn, and then the network is tested with another data set to check the success of training. One important point to remember is
that the network, if trained properly, will recognize and correctly classify only those cases included in the training set. Any new conditions not included in the training set will be
misclassified or not recognized. Feed forward, fully connected perceptron artificial neural networks (ANN), Learning Vector Quantization (LVQ), Probabilistic Neural
Networks (PNN), and Radial Basis Function Networks (RBF) are some of the available networks.  Each of the methods has its advantages and limitations. 

Computational Procedures
We will investigate several methods of analytic trace computation. They will essentially give the same result. However, one must keep in mind that the Hilbert transform is only valid
for band limited data. For example, the Hilbert transform of a spike is the Hilbert transform filter itself and it is infinitely long with decay of 1/t. This is contrary to the definition of a
spike. For this reason, we prefer to use band pass filter shaped, Hilbert transform time domain filters, which will be described later.

Frequency Domain Computation


Since the real and imaginary parts of the Analytic trace are Hilbert transform pairs, then their Fourier transforms have to be causal, their amplitude spectra be the same and their
phase spectra be 90 degrees out of phase. We can therefore, form the Analytic trace by the following simple steps;

a)Transfer the seismic trace to a complex array and place it into the real part, leaving the imaginary part equal to zero.

b)Compute the Fourier transform by FFT,

c)Zero out negative frequency components, double the positive side, but leave zero and folding frequency components as are. This will create the causal Fourier transform of the
Analytic trace.

d)The inverse Fourier transform will give an input trace that is unaltered in the real part and the imaginary part of the output will contain the Hilbert transform of the input trace.

The frequency domain method preserves the original spectrum while generating the quadrature trace. Since almost all computers have standard fast FFT routines, this method
represents the fastest and most convenient procedure.

Discrete Time Domain Computation


The discrete Hilbert transform filter in the Time domain is an infinitely long antimetric filter with zero weights at the center and at all even numbered samples. Its odd numbered
samples have weights of 1/n. The positive side + signs and the negative side has - signs. (Clearbout, 1976) This filter is, in theory, infinitely long, but in practice we use a limited
length filter, which causes the spectrum of the computed imaginary part to differ from the real part. The main difficulty comes from phase discontinuities at zero and at the folding
frequencies, which cannot be duplicated by finite length filters. In order to alleviate this, we are using more convenient band -pass type filters, such as Butterworth or my own design
exponential type filters, as the Kernel functions. We design one filter to generate the real part, which is a zero phase band pass filter. The other, a 90 -degree phase rotating filter with
identical band-pass characteristics, is used to generate the imaginary part. This insures the identical amplitude spectra of both parts and excessive distortions around zero and
folding frequencies are avoided.

Gabor-Morlet Decomposition
One of the problems associated with the Hilbert transform is that it is only valid for narrow band signals. We will have an increasing degree of uncertainty with increasing bandwidth.
For example, in the extreme case, the spike is the widest bandwidth signal and its Hilbert transform is the time domain response of the transform, which is infinitely long. This is
contrary to the locality of the spike. Luckily our data is somewhat band limited by its very nature, so we can use the Hilbert transform to some degree. In Gabor -Morlet
decomposition we divide the signal band of the original data into smaller Gabor-Morlet bands. Gabor-Morlet filters are exponentially weighted complex cosine wavelets:

                                                           (2.1)

Decomposition is done by convolving the data by a series of Gabor-Morlet wavelets generated for a sequential series of values. Gabor wavelets were first introduced to seismic
processing by Morlet et al (1982). I am enclosing a discussion of Gabor-Morlet decomposition by Koehler at the end of this report. Since the wavelets are complex valued and
Analytic, their output will also be Analytic, complex valued. These sub-bands are summed to form the real and imaginary parts of the wide-band Analytic trace. The sub-bands are
generated equally spaced in the octave scale; hence they do not cover the zero frequency vicinity. Usually 7-21 sub-bands are sufficient. We use this decomposition in the spectral
balancing program `SBAL'.
Simple Harmonic Motion Method
This method was our original approach, since it gave us the initial reasoning for generating the complex trace.  We have considered the geophone as a recording device, which
produces electrical current proportional to the velocity of the seismic waves arriving in the vicinity of its implantation.  Therefore, the recorded data is proportional to the kinetic portion
of the energy flux at the surface.  Koehler suggested that if the seismic trace were integrated, then we would have a trace somewhat similar to the trace that we would have recorded
if we could measure the position of the geophone; thus a trace corresponding to the potential portion of the energy. Since the total energy is the same in both components, we have
to equalize the running sum of squares over some long time window. We have found that due to this long equalization window and the effect of integration, the results are not
spectrally well balanced. For this reason we have elected not to use this method.

Computation of Basic Attributes


In this section we will cover all of the attributes computed directly from individual traces, without any additional traces in the form of CDP gathers, common shot traces or a stacked
section. We will cover attributes requiring more than one trace at a time in the Geometrical Attribute section. We will give the mathematical formulation of each attribute and indicate
their direct and possible relation to the physical properties of the subsurface. Taner et. al. (1979) gave the initial formulation of seismic attributes as applied to seismic data
interpretation. The paper covered five attributes, envelope amplitude, instantaneous phase, instantaneous frequency, weighted mean frequency and apparent polarity. Their
application was later discussed by Robertson and Nogami (1984) for thin bed analysis, and Robertson and Fisher (1988) for general interpretation.

In the following pages we will cover the conventional and the new attributes. We assume that the data going into the attribute computation have been adequately
processed to contain mainly the subsurface reflection characteristics. It is important to point out that, we highly recommend the use of 32 bit floating point seismic
data as input. 8 bit fixed point data, used regularly for visual interpretation, do not contain sufficient dynamic range (+/-128) to produce any reliable results. Some of
the recommended processing steps have been discussed in a Rock Solid Images technical paper "Seismic Data Processing with Post-Stack Attribute Computation as
a Consideration" by: Jago and Taner.

We further assume that the trace and its quadrature (Hilbert transformed component) have been computed previously, using the original data trace for Class I and the signal to noise
improved trace for Class II attributes. The rest of the attributes are computed from this complex trace. Complex trace generation methods are covered elsewhere in this report.

Envelope
Let the Analytic trace be given by:

                                                                            (3.1)

'>Where f(t) is the real part corresponding to the recorded seismic data and g(t), the imaginary part of the complex trace, is the Hilbert transform of f(t). Then the envelope is the
modulus of the of the complex function;

                                                                (3.2)

E(t) represents the total instantaneous energy and its magnitude is of the same order as that of the input traces. It varies approximately between 0 and the maximum amplitude of
the trace. As indicated on equation (3.2), the envelope is independent of the phase and it relates directly to the acoustic impedance contrasts. It may represent the individual
interface contrast or, more likely, the combined response of several interfaces, depending on the seismic bandwidth. Trace envelope is a physical attribute and it can be used as an
effective discriminator for the following characteristics:

·Represents mainly the acoustic impedance contrast, hence reflectivity,

·Bright spots,

·Possible gas accumulation,

·Sequence boundaries,

·Thin-bed tuning effects

·Unconformities,

·Major changes of lithology,

·Major changes in depositional environment,

·Lateral changes indicating faulting,

·Spatial correlation to porosity and other lithologic variations,

·Indicates the group, rather than phase component of the seismic wave propagation,

Rate of Change of Envelope


Time rate of change of the envelope shows the variation of the energy of the reflected events. The derivative computed at the onset of the events shows absorption effects. A slower
rise indicates larger absorption effects. We use a specially designed time domain filter to perform the differentiation.  

                                                                    (3.3)
Where * denotes convolution, and diff(t) is the differentiation operator.

Events with a sharp relative rise also imply a wider bandwidth, hence less absorption effects. This attribute is also a physical attribute and it can be used to detect possible fracturing
and absorption effects.

·Sharpness of the rise time relates to absorption in an inversely proportional manner,

·It is affected by the slope, rather than envelope magnitude,

·Lateral variation shows discontinuities,

·It is used in the computation of group propagation direction. When compared with phase propagation direction, it may indicate dispersive waves.

Second Derivative of Envelope


The second derivative of the envelope gives a measure of sharpness of the envelope peak, which may be more useful as a principal attribute display. In black and white displays, it
shows all peaks of the envelope, this corresponds to the all of the reflecting interfaces visible within the seismic bandwidth.

·Shows all reflecting interfaces visible within seismic band-width,

·Shows sharpness of events,

·Indicates sharp changes of lithology,

·Large changes of the depositional environment, even when the corresponding envelope amplitude may be low.

·Very good presentation of the image of the subsurface within the seismic bandwidth.

Instantaneous Phase
The argument of the complex function is the instantaneous phase:

                                                                  (3.4)

We display instantaneous frequency in degrees and use the color wheel to display the phase continuously. Zero degree is displayed as yellow, +120 degrees is displayed as red
(magenta) and -120 degrees (+240 degrees) displayed as blue (cyan). All phase angles between these are color interpolated. In the default color selection we are using 64 different
colors on the color wheel. This gives a rather smooth color variation in the phase display. In a number of displays we have found that 8 different colors are sufficient. These colors
represent 45-degree phase increments. The phase information is independent of trace amplitudes and it relates to the propagation phase of the seismic wave front. Since, most of
the time, wave fronts are defined as lines of constant phase, the phase attribute is also a physical attribute and can be effectively used as a discriminator for geometrical shape
classifications:

·Best indicator of lateral continuity,

·Relates to the phase component of the wave-propagation.

·Can be used to compute the phase velocity,

·Has no amplitude information, hence all events are represented,

·Shows discontinuity, but may not be the best. It is better for showing continuity.

·Sequence boundaries,

·Detailed visualization of bedding configurations,

·Used in the computation of instantaneous frequency and acceleration,

Intensity Varying Phase Display


Since instantaneous phase information is independent of amplitude information all of the events, including the ones with very low amplitudes, are displayed with the same intensity.
In order to show the instantaneous phase of the more significant events, we have combined the envelope amplitude information with the phase information and it is displayed with its
own color table. The envelope amplitudes are divided into 8 different levels. Each level is represented as successively lighter shades to indicate lower envelope amplitudes. The Color
table is also divided into 8 different regions. Each region has 8 colors similar to the original phase color wheel, except here they represent the phase in 45-degree increments. This
way the phase display shows the highest amplitude events with the most intense colors and the weakest events will appear as the least intense, pastel colors.

Normalized Amplitude Trace


In some instances a black and white section with balanced amplitudes similar to those of the seismic section is required for interpretation. Conventional short window AGC
application changes the reflected event characteristics; hence the interpreters do not favor them. The display without the change of these basic characteristics is the cosine of the
instantaneous phase. This display has all the details of the instantaneous phase section and shows the desired lateral continuity. The display has no amplitude information and its
envelope is constant unity.

Instantaneous Frequency
Time rate of change of phase is the instantaneous frequency:

                                                                            (3.5)

Since the phase function is multi -valued with jumps, it is better to compute the time rate of change as the derivative of the arctangent function, which avoids the
discontinuities:

           (3.6)

Instantaneous frequency is displayed by a color table which starts with red as the lowest frequency and it gradually changes to yellow to green and finally to blue shades for higher
frequencies. The computed output is given in units of cycles per second. Instantaneous phase represents the phase of the resultant vector of individual simple harmonic motions.
While individual vectors will rotate in clockwise motion, their resultant vector may, in some instances, form a cardioid pattern and appear to turn in the opposite direction. We
interpret this as the effect of interference of two closely arriving wavelets. This can also be caused by the noise interference in the low amplitude zones. Because of these reversals,
the instantaneous frequency will have unusual magnitudes and fluctuations. Since instantaneous frequencies are influenced by the bed thickness, we would like to observe them
without too much interference. This we accomplish by using several adjacent traces to form a consistent output. It has been shown that instantaneous frequency, computed as the
time derivative of instantaneous phase, relates to the centroid of the power spectrum of the seismic wavelet.

Instantaneous frequency computation, due to its interpretational importance, has been a subject of a number of papers. O'Doherty suggests a different way to compute the
instantaneous frequency. Consider the Analytic trace F(t) and its autocorrelation function . Let the Fourier transform of the analytic trace be represented by and the
autocorrelation function be given by . Therefore, the normalized autocorrelation function time and frequency responses relate as:

                          (3.7)

The time derivative of the autocorrelation function corresponds to the multiplication of the power density spectrum by iw. Therefore, the derivative computed at the zero lag represents
the centroid of the power density spectrum of the seismic event (Note that this is not the same as the instantaneous frequency computed as the time derivative of phase):

                (3.8)

Since the derivative of the real part of the complex autocorrelation at time zero is equal to zero, the value at the first time lag represents the phase equal to the rate of change of
phase per sample. The autocorrelation function is computed over a number of samples, which represent the averaging window. This computation, therefore, will be less affected by
superimposed reflections. Unbiased mean frequency is similar to the carrier frequency of radio signals.

Instantaneous frequencies relate the wave propagation and depositional environment, hence they are physical attributes and they can be used as effective discriminators:

·Corresponds to the average frequency (centroid) of the power spectrum of the seismic wavelet.

·Seismic character correlator in lateral direction,

·Indicates the edges of low impedance thin beds,

·Hydrocarbon indicator by low frequency anomaly. This effect is some times accentuated by unconsolidated sands due to the oil content of the pores.

·Fracture zone indicator, they may appear as lower frequency zones.

·Chaotic reflection zone indicator, due to excessive scatter,

·Bed thickness indicator. Higher frequencies indicate sharp interfaces or thin shale bedding, lower frequencies indicate sand rich bedding.

·Sand/Shale ratio indicator in a clastic environment

Weighted Mean Frequency


In the paragraph above, we described a method of computation of the time averaged instantaneous frequencies from the derivative of the complex autocorrelation function. By
variation of the autocorrelation computation window, we can control the smoothing window. The shortest possible time window, one sample, should give results similar to the
derivative of the arctangent formulation. A second method of smoothing is to use the envelope, or its square (the total power) as a weighting function for low pass filtered
instantaneous frequency:

                                                                      (3.9)
Where T is the smoothing time window.

This frequency attribute is less influenced by short wavelength effects. Longer wavelength factors, such as absorption due to thick beds or massive sand bodies, will change the
propagating wavelet characteristics that can be observed on the weighted mean frequency attribute. It is a physical attribute, indicating longer wavelength variations.

Thin Bed Indicator


One piece of information we can extract is the locations where instantaneous frequencies jump or go in the reverse direction. As we have discussed earlier, these jumps are
indicative of closely arriving reflected wavelets. Therefore, the time derivative of the phase function will contain the indicators for thin beds, in the form of large variations of
instantaneous frequency. Its smooth variation will relate to the bedding characteristics, which we will have to investigate further. The thin bed indicator is, therefore, computed as the
difference between the instantaneous and the time-averaged frequencies:

                                                                           (3.10)

This attribute shows the interference zones in phase. It is a physical attribute since it relates to closely spaced events. It can be used in detailed studies:

·Computed from large spikes of instantaneous frequency, indicates overlapped events

·Indicates thin beds, when laterally continuous,

·Indicates non-reflecting zone, when it appears laterally at random, like ‘salt and pepper’,

·Shows fine detail of bedding patterns.

Instantaneous Acceleration
The time derivative of instantaneous frequency, by definition, gives the instantaneous acceleration. We can compute this both from instantaneous frequency and from time averaged
instantaneous frequency. It is obvious that the time derivative of instantaneous frequency will accentuate the local frequency jumps. Consequently, it will make the thin bed indicators
more prominent. It should also indicate, to some degree, the effect of absorption by showing the frequency dispersion of seismic signals going through unconsolidated or quickly
deposited layers.

                                                                                      (3.11)

The acceleration computation can also be made by the O'Doherty method. The second derivative of the complex autocorrelation function corresponds to the multiplication of the
power density spectrum by the square of w. Therefore, the second derivative computed at zero lag divided by the modulus of the autocorrelation function at the zero lag will be equal
to the acceleration;

        (3.12)

It is interesting to note that this equation is same as equation 3.15 given below. Equation 3.15 gives the square of the RMS frequency (second moment of the power density
spectrum). Here (equation 3.12) we compute the instantaneous acceleration. The derivative of the time averaged frequency will be subtler, however, these displays need further
investigation.

·Accentuates bedding differences,

·Higher resolution, may have somewhat higher noise level due to differentiation,

·May have some relation to elastic properties of beds.

Instantaneous Band Width


Barnes (1992) and O'Doherty (1992) have shown there are three attributes that relate to one another in a geometrical or vectorial manner, similar to the statistics of observations. The
frequency corresponding to the centroid of the power spectrum of a wavelet is (we shall call it the Average Frequency):

                                                                                           (3.13)

The variance with respect to the centroid frequency is given by:

                                                                                    (3.14)
And RMS frequency (the second moment of the Power spectrum) is given by the expression:

                                                                                    (3.15)

By expanding the variance equation, we could easily show that

                                                                                                   (3.16)

We can now examine these statistical measurements of the power spectrum in the form of useful attributes. These computations represent the statistics of the seismic wavelet
computed over some time window. Hence, they are more closely associated with the time smoothed instantaneous attributes. We will, however, compute and display them
continuously for all of the data samples. By definition , the centroid frequency is the mean frequency where an equal amount of energy exists on either side of this frequency.

The variance with respect to the mean frequency (standard deviation) indicates the width of power spectral density distribution over a band of frequencies; hence we can use it as an
indication of the spectral bandwidth. Barnes (1992) suggests instantaneous bandwidth can be computed by:

                                              (3.17)

Where denv(t)/dt is the time derivative of the envelope.

This equation measures the absolute value of the rate of change of envelope amplitude. We could also compute the instantaneous bandwidth from the geometric equation shown
above (equation 3.16) using centroid and RMS frequency measurements. However, Barnes’ expression is simple enough to be practical. We compute and display the bandwidth in
terms of octaves.

Instantaneous bandwidth is a statistical measure of the seismic wavelet, but it relates to various physical conditions:

·Represents seismic data bandwidth sample by sample. It is one of the high-resolution character correlators.

·Shows overall effects of absorption and seismic character changes.

Instantaneous Dominant Frequency


The RMS frequency of the power density spectrum represents a biased mean towards the dominant frequency band. Equation 3.14 gives a general idea of the computation.
Following O'Doherty's reasoning we can show that the second derivative of the normalized complex autocorrelation function will give the required results. However, we use a second
method of computation as suggested by equation 3.16. Since instantaneous frequency, computed as the time derivative of instantaneous phase, represent the mean frequency
(centroid of the power spectrum), then the centroid of the second moment of the power spectrum, or the RMS frequency, is obtained by:

                                                                                                 (3.18)

Barnes calls this the dominant frequency. The display for this is similar to the instantaneous frequency display, in units of cycles per second.

Instantaneous Q Quality Factor


Barnes also suggests (in reference to definitions given by Johnston and Toks öz, 1981) that the instantaneous quality factor q(t) can be defined by the expression:

                                                                     (3.18)

Where decay(t) is the instantaneous decay rate, which is defined as the derivative of the instantaneous envelope divided by the envelope.

Except for a factor of , this is similar to the instantaneous bandwidth. The decay rate can take both positive and negative values. Hence, the instantaneous quality factor is the
ratio of instantaneous frequency to twice the instantaneous bandwidth. Barnes points out that this definition is consistent with the standard definitions of the quality factor (Close,
1966 and Johnson and Toksöz, 1981). We must point out that this Q computation is the short wavelength variation of the Q value, hence it gives relative values. It is a transmissive
attribute, similar to the interval and instantaneous velocities. It is also a physical attribute with a strong relation to porosity, permeability and fracture.

·Indicates local variation of Q factor, similar to the relative acoustic impedance computation from the seismic trace.  Longer wavelength variation should be computed by spectral
division and added to this attribute.

·May indicate liquid content by ratioing pressure versus shear wave section Q factors.

·Indicates relative absorption characteristics of beds.

·It is a transmissive attribute; its various wavelength components should be estimated in a similar way to the average velocity and velocity inversion procedures.

Relative Acoustic Impedance


We assume that the seismic data has been processed to have minimum noise and multiple contamination and it contains zero phase, wide band wavelet illumination. Based on this
assumption the seismic trace represents the band limited reflectivity series, which can be expressed as:
                                                                                                              (3.19)

Which is:

                                                                                                        (3.20)

Therefore, by integrating the zero phase trace, we will get the band-limited estimate of the natural log of the acoustic impedance. Since it is band limited, the impedance will not have
absolute magnitudes and the stack section is usually the estimate of zero offset reflectivity; hence it is called relative acoustic impedance.

                                                                                              (3.21)

In practice, however, due to noise and imperfect spectral content of the seismic data, relative acoustic impedance computed by integration will develop arbitrary long wavelength
trends. Since the seismic data does not contain any viable information at very low frequencies (due to band pass filtering in the field and during processing) these long wavelength
trends cannot be utilized. We remove these with a low cut filter.

Computation is a simple integration followed by a low pass filtering, without any exhaustive inversion. It reflects physical property contrast, hence it is a physical attribute effectively
utilized in many calibration procedures.

·Relative acoustic impedance shows band limited apparent acoustic impedance contrast,

·It relates to porosity

·High contrast indicates possible sequence boundaries,

·Indicates unconformity surfaces,

·Indicates discontinuities.

Wavelet Attributes  (Principal or Response Attributes)
Wavelet (characteristic or response) attributes are, by definition, the instantaneous attributes computed at the maximum points of the trace envelope. Intuitively, the maximum of the
envelope represents a position where the majority of the energy from different frequency bands is in phase. Therefore, attributes computed at that position would have some direct
statistical implication to the wavelet’s spectral characteristics. A number of authors (Bodine, 1984, Barnes, 1991) have published studies of the statistics of the analytic signal and
the relationship between the Fourier transform and the instantaneous attributes. The statistics show that the attributes computed at the envelope peak relate directly to the various
moments of the power density spectrum. Therefore computation and display of these attributes will provide an insight into the reflected wavelet characteristics. Bodine (1984) calls
these attributes the `response' attributes. I prefer the `Wavelet' attributes, because they summarize the wavelet characteristics. Since only one attribute value is to be displayed over
the length of a wavelet, we have to determine and define the wavelet length.

Definition: We assume that a seismic wavelet, or a compound wavelet, occupies the time span between two adjacent envelope minima. In reality, wavelets will extend beyond these
minima, however, for display purposes we consider the minima as the boundary. Furthermore the part between the minima represents a significant portion of the wavelet.

Therefore, we search for the envelope minima and maxima positions. All envelope maxima positions are used to obtain the instantaneous attributes and these values are assigned
over the time zone between the minima on either side of the maximum position. It has been found, however, that the computation of the attributes at the sample points is not
sufficiently accurate. For example, if we consider 4 millisecond -sampled data, for a 40 Hz. wavelet each sample interval will have a 60 -degree phase difference. This is much too
large to be useful. In order to be accurate, therefore, we compute the position of the maximum of the envelope by quadratic interpolation. All of the other wavelet attributes are also
computed at the interpolated maximum position of the envelope.

The following are computed and displayed as the wavelet attributes:

·Wavelet envelope

·Maximum positive time derivative of wavelet envelope. This attribute is computed at the onset of the wavelet at the position where the time derivative is a maximum. It is obvious the
time derivative at the envelope maximum will be zero, hence useless. The maximum time derivative scaled by the envelope maximum will have implications of absorption.

·Wavelet second derivative of envelope. This indicates the sharpness of the peak, hence some indication of bandwidth.

·Wavelet phase. Lateral similarity is an excellent indication of depositional continuity.

·Apparent polarity. Polarity is assigned based on the Wavelet phase. If the phase is between -90 and +90 degrees a positive polarity is assigned, otherwise a negative polarity is
assigned. Its magnitude is equal to the interpolated envelope magnitude.

·Wavelet instantaneous frequency. Similar to phase, a good indicator of lateral continuity and bandwidth.

·Wavelet acceleration

·Wavelet bandwidth

·Wavelet dominant (RMS) frequency. This is direct computation of the RMS frequency of the wavelet power spectrum.

·Wavelet Q quality factor


Wavelet attributes use the same color table as the instantaneous attributes in order keep their relationships apparent.

All of these attributes have similar discriminatory significance, but they relate to individual events, rather than to individual samples.

Geometrical Attributes
Introduction
The phase component of seismic data contains an expanse of useful information. This information can be obtained if the seismic trace is considered either as a complex entity, or
as an analytic function consisting of the recording of the potential and the kinetic components of the energy flux at the surface of the earth. While the measurement of the phase and
its time derivative gives direct information to the state and variation of energy in a temporal sense, the measurement, extended to include the spatial information, yields information
on the wave number and the visible direction of propagation. It is well known that, wave propagate in two separate mode, phase and group. In dispersive medium these two modes
will have different propagation velocities. Here, we assume a non-dispersive mode and consider the phase component only. Separate computation of phase and group propagations
have been discussed by Barnes (1994).

A further, and perhaps more important benefit, comes from the redundant information contained along the wave front which helps to reduce the effects of noise. One of the more
useful attributes, the instantaneous frequency, suffers from the influence of noise, which breaks the spatial continuity. This appears, even in time averaged frequency displays, as
trace wide color streaks. O'Doherty (1992) has shown that by including the spatial information in the averaging process, the instantaneous frequency and dip computations can be
stabilized to the degree that the events can be tracked with more confidence.

Let f(t,x) be the recorded seismic trace which is the real part of the complex trace, then  g(t,x) is computed by Hilbert or Gabor filtering of the  f(t,x) trace, then the complex traces
are defined by (now in two dimensional sense)

                                                           (4.1)

Instantaneous amplitude is given by:

                                                  (4.2)

Instantaneous phase is given by:

                                             (4.3)

Instantaneous frequency is given by:

                                                                  (4.4)

One of the main problems is the discontinuity in the phase function. It is not a continuously increasing function. This gives rise to negative instantaneous frequencies, most likely
due to the interference caused by two or more overlapped wavelets. The main hypothesis is ``Simple seismic wavelets are such that their amplitudes rise to a maximum at a rate
proportional to their general spectral content and their phase increases continuously without any local reversal''. It has been shown that the rise time is affected by the energy loss at
higher frequencies or is due to dispersion effects. One method of Q computation is based on the wavelet envelope rise time. According to our proposed condition of continuous
phase increase, we consider the phase reversals are due to the interference of closely spaced wavelets with opposite polarities. If we plot the hodogram (the trace vector in the
complex domain) the end of the vector rotates in a counter clockwise direction, in a somewhat circular orbit with a time varying radius. The interference appears as a cardioid pattern
that shows the resultant vector going in the opposite direction for a short time period and then proceeding in the original direction. The average angular velocity of the vector is the
mean frequency of the corresponding wavelets. This is similar to the carrier frequencies of radio waves. In our computation we will have to compute and remove the slowly time
varying component of the angular velocity (instantaneous frequency) or the mean slope of the phase function. The remainder will show the local deviations. These local deviations,
when displayed in a sectional form, will show the thin bed structure of the seismic data. Local deviations can also be due to the influence of the noise vector, therefore, we have to
separate the effects of the noise from those of the wavelet interference. Usually the noise effects are visible in the zones of low signal amplitude. Interference occurs at the arrival of a
second signal with larger envelope magnitudes. This condition has led us to develop a weighted filtering scheme for smoothed instantaneous frequency determination.

This weighted frequency is a slowly varying function of time and represents, essentially, the Wavelet frequency. Since the reflectivity function of the earth is finely structured and the
wavelets used to illuminate it have, relatively, a much longer wavelength, we are essentially observing the interference patterns, rather than a direct observation of images. In the
future we will need to develop some new methods to take advantage of the interference patterns and make better and finer detailed estimates of the subsurface images. Thin bed
indicators, instantaneous and average frequencies were covered in the second section.

Instantaneous frequency was one of the earliest used attributes. It is the rate of change of the instantaneous phase with respect to time, and is a temporal measurement. Similarly,
spatial measurements lead to the wave numbers, which will be useful in indicating the direction of the phase component of the propagating wave field. The Eikonal equation gives us
the relationships between the temporal and spatial variables.

Let be the wave -number (radians/meter) in the x direction and the wave number in the vertical z direction and  be the angular velocity (radians per second) and V the
propagation velocity (meters/second), then a plane harmonic wave represented in the form:

                                                        (4.5)

will satisfy the scalar wave equation. Substituting this into the wave equation we will obtain the Eikonal relationship:

                                                                                           (4.6)

This can also be written as:


                               

                                                                                           (4.7)

The time gradient on the seismic section is , then from Eikonal equation, this is equal to:

                                                                                   (4.8)

which is the ratio of the horizontal wave number to temporal frequency.

In the case of finely sampled data sets ' and ` ' can be computed directly from the lateral and temporal derivatives of the instantaneous phase function. If the data is coarsely
sampled we will be subjected to spatial aliasing, hence we cannot obtain a good measurement of dips this way. In these cases local dip scanning gives a little coarser but more
accurate results. An alternative way is to interpolate the data to finer spatial sampling intervals and then compute the dips. However, this is not an efficient method because, during
the interpolation, there is an implied step of dip estimation in order to interpolate the aliased data properly.

Ron O'Doherty showed that the ' method of computing dips suffers in the same way as the anomalous variations in the instantaneous frequencies. For the same reason he
recommended temporal and spatial smoothing in order to get sectional consistent dips. We are using amplitude weighted smoothing similar to the frequency smoothing in order to
minimize the effects of the low amplitude noise.

O'Doherty proposes a slightly different method of smoothing, which is done during the computation of the instantaneous frequencies and wave numbers.

It was shown by a number of authors that the instantaneous frequency computed at the maximum of a non -complicated wavelet (like zero or nearly zero phase wavelets)
corresponds to the centroidal frequency of the power spectrum of the wavelet:

                                                                          (4.9)

We know that the autocorrelation function is the inverse Fourier transform of the Power density spectrum:

                                                        (4.10)

And the zero lag of the auto-correlation function is given by:

                                                                              (4.11)

Which is equal to the denominator of the above expression 4.9. On the other hand, the derivative of the autocorrelation function in the frequency domain is its Fourier transform
multiplied by :

                                       (4.12)

Which is essentially the numerator of the equation 4.9. At the zero lag the exponential portion of the above expression vanishes to give us the derivative of the complex
autocorrelation function at the zero lag. We can therefore compute the instantaneous frequency by the ratio of:

                                                                          (4.13)

In order for this equation to make sense, must be uniquely defined at all frequencies and must be a complex valued function. And, since it is a complex autocorrelation
function, it should have Hermitian type symmetry - positive lags are the complex conjugate of the negative lags. Its first derivative at the zero lag will be purely imaginary. The
imaginary part of the autocorrelation function will be zero and the first derivative of its real part will also be zero. Therefore, the instantaneous phase of the complex autocorrelation
function at time lag one will represent the amount of phase change per sample time unit. So the instantaneous frequency, defined as radians per sample, will be given by:

                                                                                         (4.15)

Since instantaneous frequency is the time derivative of the instantaneous phase, equation 4.13 gives the evaluation of this derivative at the peak of the envelope, which is described
as the mean frequency:
                                                                                       (4.16)

It is interesting to note here that, since the autocorrelation function is complex valued, we will have 4 samples per cycle at the folding frequency. In order to compute the
instantaneous frequency correctly we have to compute the phase as shown in equation 4.14 from zero to 180 degrees, not between -90 to +90.

O'Doherty suggests using time windows long enough to suppress the effects of the noise. A time window of the order of an average period should be sufficient. This computation
should produce instantaneous frequencies similar to the weighted mean frequencies described above. In the computation of horizontal wave numbers however, this method may be
more convenient, since it uses less samples.

O'Doherty also suggests spatial averaging of the autocorrelation function as well as the temporal averaging. This way the influence of noise is minimized and laterally coherent
wavelet information is strengthened.

The second derivative of phase with respect to the spatial variable x shows lateral (linear) continuity. If the second derivative is zero or near zero then the lateral change of phase is
linear, hence the event is linearly continuous. Any large second differences indicate some form of discontinuity or curvature of the arrival times of the event. For the sake of
practicality we will use linear continuity over 3 adjacent traces as the measure of lateral continuity.

However, before we consider the lateral continuity, we will look into the other lateral measurement, the wave number in the spatial x direction. The wave number is computed in the
same manner as in the temporal direction. We have found, as mentioned above, that O'Doherty ’s method, using the first derivative of the autocorrelation function in spatial
direction, has advantages and it uses less spatial samples. Therefore the wave number is given by:

                                                                                      (4.17)

Where is the first space lag of the autocorrelation function computed in the lateral direction. The number of traces that are included in the autocorrelation computation controls
the degree of smoothing.

According to the Eikonal equation, the ratio of temporal frequency to the spatial wave number is the time gradient of the wave front arriving at the surface.

                                                                                 (4.18)

And,

                                                                               (4.19)

Where is surface arrival angle of the wave front and is the velocity at the layer where the receivers are implanted. Scheuer and Oldenburg (1988) have used this ratio to
determine the local phase and the apparent velocity from common shot data. This measurement is displayed as the apparent dip. By assigning the surface velocity, the apparent dip
can be converted into the exit angle of the surface arriving wave front.

                                                                         (4.20)

This computation can easily be extended to 3-D data sets. Similar to the expressions given above, by introducing the component in the y direction, we will have:

                                                                              (4.21)

And;

                                                                             (4.22)

Equations 4.19 and 4.22 give the time gradients measured in x and y directions. Then the maximum gradient is given by:

                              (4.23)

And its azimuth will be:

                                                                           (4.24)

One of the important stratigraphic measurements is the lateral continuity of the reflecting horizons (Sangree, 1988 and Vail, 1988). We are also interested in the concordance or the
consistency of the bed thickness (if they are parallel or not). As we have discussed earlier, we will measure the linear continuity by the second order partial differences of the phase
in the lateral direction. Increasing values of second difference will indicate an increasing amount of reflection curvature. The direction of continuity will be obtained from the dip
measurements. A smoothed version of the lateral continuity measurement can be obtained by the inclusion of more samples in the spatial and temporal directions.  This is 
discussed in the Dip Scan section.

Another important indicator is the location of event terminations in the form of on-lap, top-lap truncation and etc. (Sangree and Vail). These event terminations are used to determine
the stratigraphic sequence boundaries. Determination of abrupt phase changes, as in the case of faults on a migrated section, is simple. The main difficulty is locating gentle dip
differences that might be associated with down-lap or top-lap conditions. In un-migrated data, faults have diffraction hyperbolae, which makes them appear continuous. Migration
process focuses these diffraction hyperbolae to their origination points, thus making the discontinuities more visible. At present, we will use the computed lateral semblance to
indicate the lateral event termination and continuity. We will also consider sequence boundary detection as one of the principal research subjects for the near future.

One way to indicate event termination is to highlight the opposite of continuity as discontinuity. That is, we will highlight the larger values of the lateral second difference of
instantaneous phase. This method will probably be less diagnostic for the beds that merge gradually. A second method is to laterally track lines of constant phase and to indicate
their terminations. This tracking could also be carried out on narrow frequency band filtered data. We would have to observe the consistency of truncation over several bands in order
to have higher redundancy. This method will be adversely influenced by noise and by zones of chaotic reflection. Therefore a simple method to check consistency will be necessary.
Computation of Geometrical Attributes
There are several methods to compute geometrical attributes. Earlier in this chapter we have discussed the method based on O'Doherty's method of smoothing in spatial and
temporal directions. A second method is to compute differences in spatial direction sample by sample. If the dips extend beyond one sample per trace this method will give aliased
results. A third method of dip scanning will overcome this difficulty. In this section we will continue to discuss the last two methods.

Event Continuity
This is an intermediate result of a hybrid attribute computation. The objective of this attribute is to develop the lateral continuity of peaks and troughs, and to classify the type of
discontinuity. At this time all real trace positive peaks are output as +1. and negative peaks output as -1. Everything else will be output as zero. The display therefore will show only
the peaks and troughs, all with same magnitude. The development phase will connect like peaks and establish discontinuities.  Based on the type of discontinuity, it will attempt to 
classify the terminations as top-lap, on-lap etc.  

Instantaneous (Phase) Dip


Instantaneous phase for three adjacent traces is computed. A parabolic curve is fitted to the phase values laterally, and the dip at the center trace is computed. This dip is then
adjusted according to the trace interval. In 3-D data sets, phase dips in in-line and cross-line directions are computed, from which the maximum dip direction and its azimuth are
obtained.  Let dt/dx and dt/dy be the corrected dips in in-line and cross-line directions respectively, then (similar to equations 4.23 and 4.24):

        

Will give the maximum dip, and its direction (Instantaneous Dip Azimuth) will be:

As mentioned above, this computation is valid for dips of up to 180-degree (exclusive) phase differences. Actual dips beyond this will produce aliased results, hence they will be
misleading. In such cases dip scanning with a greater number of traces becomes necessary. As a side note, when time migrated sections are used, it is highly recommended that
the image trace spacing should be designed to eliminate any possible aliasing problems. In most instances, conventional surface recording intervals of 12.5 or 25 meters will be
sufficient to generate images at target levels below one second at less than half of the surface interval, without violating any sampling laws.

Instantaneous Lateral Continuity


The second difference of the instantaneous phase function is computed in in -line and cross-line directions. These represent interface curvature in each direction. The maximum
curvature is assumed to be in the same direction as the maximum instantaneous dip. Maximum curvature is estimated similar to the maximum dip:

                                             (4.25)

Linearly continuous events will give zero curvature. Beds with a hummocky appearance will have non-zero curvature values. Non-reflecting zones will have highly variable curvature
values in time and space. This attribute highlights the zones of large lateral dip variation; hence it can be a good indicator of faults and fractures.

Attributes Computed by Dip Scanning


Seiscom and Digicon independently developed the Dip Scanning process in the late 1960's under the commercial names of Seis-scan and Digi-stack, respectively. The objective of
the scanning was to determine the dip of maximum lateral continuity, and use that dip to form an average data sample to produce a trace with improved signal to noise ratio output.
To measure the similarity, semblance, which was developed in 1965 by Seiscom, was used. This information was also used in the velocity analysis to provide the surface arrival dip
as part of the stacking velocity estimate. These estimates were later used for dip, depth and interval velocity computations. Since discontinuities are the inverse of continuities, dip
scan sections have highlighted discontinuity as well as continuity. In the late 1980's, lateral maximum semblance values were introduced as additional geometrical attributes.

Maximum Lateral Semblance and corresponding Dips as Lateral Continuity Attributes


The objective is to compute the maximum lateral semblance and corresponding dips. This is performed in the DIPSCN Dip scan subroutine. The subroutine accepts a user requested
number of traces as input. The center trace is assumed to be the center of scanning. All traces are assigned positive and negative distances according to their location with respect
to the center trace. Actual trace distance in in-line and cross-line are considered. This will provide the consistency of dip measurements with respect to per unit distance. The traces
are summed along the linear alignments governed by the submitted dip ranges. The semblance values are computed for each scanned dip over a Gaussian window in the time
direction. The semblance value and the dip corresponding to the maximum semblance for each data sample are saved as output. The semblance maxima are detected by parabolic
interpretation of the semblance values of three adjacent scanned dips. These semblance maxima are compared against the overall maximum semblance. If the local maximum is
larger, then the maximum semblance is updated, and part of the summed trace and the corresponding dip values are transferred to the output trace and dip trace respectively. After
all of the dips have been scanned, the maximum semblance contains the semblance value for the laterally most coherent dip within the given dip range, the dip trace contains the
corresponding dips and the output trace contains the summation of traces along the dip of maximum coherency. This last output will contain laterally most continuous events. All
three output traces are stored as attributes. The Class II attributes are computed from these lateral coherency improved output traces.

In some instances, to further improve the lateral continuity of the traces, the scanned output is scaled by a percentage of the semblance value. Since the semblance values lie
between zero (corresponding to complete lateral dissimilarity) and 1.0 (100 percent lateral continuity), multiplying with semblance values will enhance further the laterally continuous
events while suppressing zones of discontinuity. However, this additional process produces rather strong effects, hence it should be used with care.

The semblance is a measure of coherent power existing between a number of traces versus the total power of all traces, as given by;

               (4.26)

Where is the m'th  trace of the gather, and  N is the length of computation window.


The lateral "coherency" indication is computed as semblance that is determined by considering a user-selected number of traces. The number of traces depends on the signal to
noise ratio (poor S/N, more traces), bed curvature (more curvature, less traces), and higher lateral discontinuity resolution (better resolution, less traces). The semblance is generally
computed over a running 40 - to - 80-millisecond time window (shorter time windows are appropriate for wide band data). For good S/N data, 3 adjacent traces for scanning will give
higher resolution. 5 or higher numbers of traces will tend to lose lateral discontinuities and start producing longer linear components of existing events. The dip scan range and dip
increments are input in milliseconds per trace, regardless of inline and cross line distances. However, if the actual inline and cross -line spacing has been entered, then the dip
attribute output will be correctly scaled in milliseconds per 100 meters.

In 3-D data sets, dip scanning is done for in -line and cross-line traces separately. Each scan will output S/N improved trace, maximum semblance and dip. These are combined to
produce one dip, maximum semblance and S/N improved trace. Computation of maximum dip and its azimuth is given earlier in this report.

In 2-D cases the output from subroutine DIPSCN is used as direct input for Class II attribute computation. In 3-D cases the output from in-line and cross-line scans are combined by
semblance weighting as:

            

                                                                                                                                     (4.26)

Where and are in-line and cross-line semblance scanned trace outputs, respectively. Equation 4.26 shows that the final trace is generated by
proportional summation of two traces weighted by their percent semblance values.  This output is used as the input trace for Class II computation in 3-D data sets.

Dip of Maximum Lateral Semblance


In 2-D computations output from DIPSCN is directly used. In 3-D cases, dips in the in-line and cross-line directions are vectorially summed, as in equations 4.23 and 4.24, to
determine the maximum dip and its azimuth.

Smoothed dips of maximum lateral semblance


Computed dip values are low pass filtered to generate an unbiased running average estimate. This is an iterative process designed to minimize the effect of local outliers and spikes.
This is done several times using a weighted low pass filter. The weights are inversely proportional of the input raw data to the computed average. This will reduce the effect of outliers
and will give a more robust estimate of the mean, close to the mode. This attribute shows the prevailing dips in a particular depositional setting

Let D(t) be the average dip, d(t) instantaneous dips and w(i) are the weights, the average dip is computed by:

                                      (4.27)

And the weights are computed iteratively after each pass by:

                                                              (4.28)

Where K is a constant controlling the severity of rejecting outliers. Normally K=4 could be used.

Dip Variance (Local Dip versus average Dip)


This attribute measures the difference between the instantaneous dip of maximum lateral semblance and the time average dip. Time averaging is computed with a longer user
selected time window. The averaging is an iterative process to estimate the average closer to the mode without the effects of the outliers. This attribute shows the anomalous dips
within a uniformly layered environment. It could be used to detect small faults and permeability barriers.

Lateral Continuity
The maximum lateral semblance as output from DIPSCN is used directly in 2-D cases. In 3-D cases the output from in-line and cross-line scans are vectorially summed to produce
the lateral continuity attribute. This attribute shows the regions of lateral consistency, or the similarity of depositional environment. It will also show the areas where the continuity
has been disturbed.

Smoothed Maximum Lateral Semblance


Semblance values are low pass filtered to generate an unbiased running average estimate. This attribute is used in locating zones of uniform depositional settings. It is also used in
computation of hybrid attributes.

Variance of maximum coherency, instantaneous versus average


This attribute is the difference between the instantaneous maximum lateral semblance and smoothed maximum lateral semblance. It shows the local depositional anomalies. It may
also show the zones of noise or massive non-reflecting areas.

Hybrid Attributes
Most of the attributes discussed previously are defined as "Primitive" attributes. They basically measure a particular physical or geometrical condition or statistic. They are primitive,
but they can also be used as building blocks and, by combining them logically, can be used to detect more complicated conditions. The logic of the combination is experience and
knowledge based. For example, the fluid factor computed by the AVO analysis is a hybrid attribute based on experience. In this section I will give descriptions of several hybrid
attributes. These are our initial attempts to combine attributes with a knowledge-based logic. Since their introduction they have been shown to be effective discriminators. We will
continue in developing additional hybrid attributes as we incorporate an additional knowledge of Rock Physics, Geology and Reservoir characteristics.
Parallel Bedding Indicator
This is the standard deviation of instantaneous dips computed over the average dip computation window.

                                                            (4.29)

Zones with parallel bedding will have zero, or close to zero, variance or standard deviation. Increasing values of standard deviation will indicate increasing dip
variation within the window. This attribute, computed as dip statistics, provides a direct indication of the geometrical configuration of beds and interfaces.

Chaotic zone indicator


Chaotic zones are defined as zones with a high degree of lateral continuity and with arbitrarily varying bedding dips. This attribute is computed to emphasize the zones of chaotic
versus organized events. Lateral dips will stay consistent over zones of parallel dipping reflectors; hence they will have a small standard deviation of dip. Chaotic zones will have
higher lateral coherency but their dips will change more rapidly with time and space. Non-reflecting zones, while having a higher standard deviation of dip, will also have a lower
average lateral semblance. The attribute is computed as running standard deviation of dips weighted by the lateral semblance.

                                                           (4.30)

Zones of chaotic bedding will have high attribute values. Massive carbonate zones with incoherent noise, like weak reflections, will show lower values of chaotic attributes.

Zones of Unconformities
This is a hybrid attribute for detecting possible surfaces of unconformity. The routine first detects the peaks of the seismic trace envelope. The location of the peak of the envelope
represents the seismically visible bedding interface, a possible unconformity surface. Its magnitude is relative to the degree of impedance contrast. The next process is to compute
additional contrasts. Three samples of dips of maximum lateral semblance, instantaneous frequency, lateral semblance and variance of dip attributes are picked from both sides of
the interface. A normalized vector dot product is formed between the set of attributes on two sides. Since the vector dot product is normalized, (the components are the attributes),
its values will lie between + / - 1. +1 represents the highest similarity, indicating there is no difference between the sides from the dip, frequency and continuity points of view. A
smaller value will indicate some degree of difference. A dissimilarity factor is formed by subtracting the normalized vector dot product from 1.0, which is then multiplied by the
envelope peak magnitude and output as the indicator of the Zones of Unconformity. A larger magnitude output represents a higher degree of contrast.

Shale Indicator
This a hybrid attribute that combines several primitive attributes to detect possible shale zones in a clastic environment.  Shale beds are identified by their geometrical configuration
as; thin parallel beds with high lateral continuity. This attribute uses Instantaneous higher frequencies as the thin bed indicator, lower values of standard deviation of maximum
semblance dips as the parallel bed indicator and semblance and its variance as the lateral continuity indicator. The highest output value indicates the highest possibility of shale
occurrence. Lower magnitudes indicate the possibility of lithologies other than shale, such as sand or carbonate beds.

FILTER DESIGN
In this chapter we will present the design of several filters used in seismic data processing and Attribute computation. These will include two types of band pass filter, a Hilbert
transform filter in convolutional form and a special differentiation filter and a specific process to transform from the cepstrum domain to the time domain by a simple recursion. All of
these filters are explained in the text and the corresponding Fortran subroutines are included in the appendix.

Band Pass Filters


We use IEEE rules to define the pass and reject zones and the transition from and to the pass/reject zones. Since all filters are passive, they remove some energy from the input
data, therefore, the output is always equal to, or less than, the input data. If a filter suppresses more than 1/2 of the input power at some frequency band, this is called a reject
region. A pass region is defined as a frequency band where the suppression of the input is less than 1/2 power or -3 decibel. Therefore, the 1/2 power suppression point defines the
boundary between the pass and reject zones. The rate of change from reject zone to pass zone or vice versa is given as a roll-off rate in decibel-per-Octave (dB/Oct.). Normally used
rates are 8, 12, 18, 24, 36, 48, 60, 72, 120 dB/Oct. 8 dB/Oct being the most gentle change and 120 dB/Oct. being the sharpest change. In most processing filters we use roll-off
rates of 12 to 48 dB/Oct.

We have found that the amplitude or power spectrum of band-pass filters has to be continuous and continuously differentiable; that is smooth in the frequency domain. This
characteristic produces the least undulation of the envelope of the filter wavelet in the time domain, a most desirable property for high resolution and stratigraphic processing.
Another point is that the filters described here pass all of the frequency band to some degree. At no time do they completely suppress any frequency, as is the case with Ormsby
type filters. Therefore they can be inverted, and their inverses will have no undesirable poles in the frequency domain or corresponding reverberations in the time domain.

We define the filters by their low and high cut frequencies, or their 1/2 power points at the low and high frequency sides. For example, a filter with 8 to 60 Hz pass band and with
slopes of 18 dB/Oct. on the low side and 24 dB/Oct. on the high side means that the data will be suppressed by less than 1/2 power from 8 Hz. to 60 Hz. Outside these
frequencies, suppression will gradually increase beyond 1/2 power or -3 dB. On the low side, one octave lower than 8 Hz, at 4 Hz., the suppression will reach -21 dB. or 1/125'th
power of the input. On the high side, at 120 Hz. (one octave higher than 60 Hz.) it will reach -27 dB. Or 1/500'th of input power.

We will use and as the low and high cut frequencies, and and as the low and high frequency side roll-off rates respectively.

Butterworth Filter
Butterworth band-pass filters are composed of the multiplication of two sigmoid functions. The following computations are performed in the SRCBAND subroutine.

Low Pass Butterworth Filter


The power spectrum of the Low-Pass filter is given by:

                                                               (5.1)

Where f is the frequency given in Hz. We compute the constant N from the roll-off rate requirement, since the half power point at is automatically satisfied by eq. 5.1; If the
roll-off rate is , then by definition, the power spectrum at frequency will be minus ‘3.0103 ’ decibel down;

Let , then

,                      

Solving for N we get:

                                                       (5.2)

High Pass Butterworth Filter


Similarly for the High-Pass filter, we can state the conditions as:

                                                              (5.3)

The power ratio at 1/2 of the low cut frequency will be down from 1/2 power at frequency, then

Let , then

,                  

Solving for M we get:

                                                        (5.4)

Band Pass Butterworth Filter


The Butterworth Band-Pass filter power density spectrum is the product of the spectra of the Low-Pass and the High-Pass filters as:

                        (5.5)

Minimum Phase Butterworth Filters


Because of the non-zero spectral values, all of the Butterworth filters have their minimum phase equivalents. If the real part of the data set in one domain is the Hilbert transform of
the imaginary part, then their inverse, or forward Fourier transform, will be causal (one sided or the negative side of the transform will be all zeros). Since the log -amplitude spectra of
minimum phase filters are the Hilbert transform of their phase spectrum, then their cepstrum is one sided, the negative lags will be all zeros. We use this characteristic and compute
the inverse Fourier transform of the log amplitude spectrum to generate a two-sided cepstrum. This corresponds to the cepstrum of a zero phase wavelet. Let BB(f) be the power
spectrum of the Butterworth filter. (This could be low-pass, high-pass and band-pass) Then we compute:

                                            (5.6)

Since we have used only the real part with the imaginary part equal to zero, then the will be symmetrical. In order to form the cepstrum of the minimum phase wavelet we will
set the negative lags to zero:

   for,    
And double all values for lags larger than zero, leaving the zero lag as is:

   for,    

We now have the cepstrum of the minimum phase filter. Once the cepstrum of the minimum phase wavelet is obtained, then its time domain response is computed by a simple
recursion (Oppenheim and Schafer, 1975, p. 504). This is done in the CEPTIM subroutine. Let be the n'th element of the cepstrum and the m'th element of the filter in the time
domain, then by recursion we compute later elements of the filter from earlier computed elements as:

,  thus the first element will be     '  and

                                                          (5.7)

Recursion will produce the rest of the filter weights. In order to have gradual reduction of filter amplitudes down to zero at the tail end, we use Butterworth type
sigmoidal tapers as used in the TAPER subroutine.

Taner Filter:
We have found that Ricker wavelets have the most desirable shape; their envelope is a decaying exponential function. All of the filters designed by specifying four corners of the
spectrum, like Ormsby filters, will have envelopes with many lobes. These lobes are very disagreeable for the fine definitions required by stratigraphic processing, as well as regular
processing. In order to achieve smoothness of the wavelet envelopes, I experimented with exponential functions in the log and octave domain. Low and high pass filters are simple
exponential functions, whose characteristics are controlled by the user given band limits and roll -off rates.

Low Pass Filter


The log-amplitude spectrum of the Low-Pass filter is given by:

                                                     (5.8)

Where x is the frequency given in octaves as . If then we will have 1/2 power or in log scale. Therefore;

We compute the constant from the roll-off rate requirement, since half power point at x=xH is automatically satisfied by eq.5.8.; If the roll-off rate is , then by definition the
power spectrum at 1 octave higher than frequency will be ' decibel down;

Or,

Solving for we get:

                                         (5.9)

High Pass Filter


Similarly for the High-Pass filter, we can state the conditions as:

For we will have   , hence;

                                                                                       (5.10)

The power ratio at a frequency one octave lower than the low cut frequency will be lower than the 1/2 power at frequency. Therefore;

Or,

Solving for we get:


Solving for we get:

                               (5.11)

Band Pass Filter


The Band-Pass filter spectrum in log amplitude and octave domain is the sum of the Low-Pass and the High-Pass filters, which in turn means the products of the spectra of the Low -
Pass and the High-Pass filters.

    (5.12)

Computation of zero and minimum phase Taner filters are given in TANER subroutine.

CONVOLUTIONAL HILBERT/BUTTERWORTH FILTER


The conventional Hilbert filter in the time domain is of the form 1/n as given by Clearbout. While these filters are theoretically correct, they are not suitable for practical application
due to their length and the excessive distortions caused by the discontinuity at the high end of the frequency spectrum. Also, the Hilbert transformed trace does not have the same
amplitude spectrum as the input trace, because only the quadrature trace is computed by convolving the input trace with the Hilbert filter. In order to have the same spectrum for both
the real and imaginary parts of the complex trace, we designed Hilbert filter pairs to generate both the real and the imaginary parts. The real part of the Hilbert filter is symmetrical
and the imaginary part is anti-symmetrical. The amplitude spectrum is computed as described above in the Butterworth filter design section. Since the Hilbert filter’s real and
imaginary parts are also Hilbert pairs, then their amplitude spectrum will be causal. Since both the real and imaginary parts have the same amplitude spectrum, we double the
computed Butterworth amplitude spectrum for all positive frequency values, leaving the zero frequency as is. An inverse Fourier transform will form the complex Hilbert filter in the
time domain. In another way, a cosine transform of the amplitude spectrum will generate the symmetrical real part of the Hilbert filter and a sine transform will form the antimetric
imaginary part. The HILBAND subroutine generates the pair of Hilbert filters with a Butterworth spectrum controlled by the user’s bandwidth and roll-off rate requirements. Hilbert
filters can be designed with Taner filter type spectral controls by replacing the spectral design portion of the subroutine.

CONVOLUTIONAL DIFFERENTIATION FILTER


In the sampled time domain the ideal differentiation filter is of the form:

We design the filter weights so that the derivative at time zero will be:

                                             (5.13)

If ,  then    ,

And
,      then     ,
should result by this convolution. Therefore;

                                         (5.14)

Or;

                                                                               (5.15)

Therefore,

                       (5.16)

Therefore the exact solution will be:

                                                                                              (5.17)

And,

As we can see, this operator decays by a slow 1/t rate, making a very long filter. Secondly, it will have a very sharp edge at the folding frequency, which will create undesirable side
effects on the output. In order to minimize these side effects we have designed differentiation filters with gentler slopes on the high frequency side.

The subroutine DIFROP generates the convolutional differentiation operator.


References

© 1992-2001 Rock Solid Images, all rights reserved

You might also like