Download review of seismic reflection - University Of Nigeria Nsukka

Document related concepts

Seismic communication wikipedia , lookup

Earthquake engineering wikipedia , lookup

Seismometer wikipedia , lookup

Reflection seismology wikipedia , lookup

Seismic inversion wikipedia , lookup

Surface wave inversion wikipedia , lookup

Magnetotellurics wikipedia , lookup

Earthscope wikipedia , lookup

Transcript
1
IBEKWE, ABRAHAM IKECHUKWU
(PG/M.Sc/09/51925)
REVIEW OF SEISMIC REFLECTION
DEPARTMENT OF PHYSICS AND ASTRONOMY
FACULTY OF PHYSICAL SCIENCES
Ugwoke Oluchi C.
Digitally Signed by: Content manager’s Name
DN : CN = Weabmaster’s name
O= University of Nigeria, Nsukka
OU = Innovation Centre
2
REVIEW OF SEISMIC REFLECTION OPERATIONS
BY
IBEKWE, ABRAHAM IKECHUKWU
(PG/M.Sc/09/51925)
A PROJECT
SUBMITTED TO THE DEPARTMENT OF PHYSICS AND ASTRONOMY,
FACULTY OF PHYSICAL SCIENCES,
UNIVERSITY OF NIGERIA, NSUKKA,
IN PARTIAL FULFILMENT FOR THE AWARD OF MASTER OF SCIENCES (M.Sc)
IN SOLID EARTH GEOPHYSICS.
SUPERVISORS
PROF. F. N. OKEKE (FAS)
DR. J. U. CHUKUDEBELU
NOVEMBER, 2013
3
CERTIFICATION
This is to certify that this project was submitted by Ibekwe, A. Ikechukwu to and approved by the
Department of Physics and Astronomy, Faculty of Physical Sciences, University of Nigeria,
Nsukka, in partial fulfillment for the award of Master of Sciences (M.Sc) in Solid Earth
Geophysics.
__________________________________________
Head of Department
__________________________________________
Supervisor
__________________________________________
Supervisor
_______________
Date
______________
Date
______________
Date
4
ACKNOWLEDGEMENT
“Dreams are not the things you see when you sleep. Dreams are those things that keep you from
sleep!”
The completion of this project is a dream come true and I would like to acknowledge all those who
in one way or another contributed to this great success.
Firstly, I would like to acknowledge the Head, staff and students of the Department of Physics and
Astronomy, UNN, for the knowledge and time we shared. Your inputs actually contributed to my
success. Most especially my supervisors, Prof (Mrs.) F. N. Okeke (the most compatible person I
have ever worked with) and Dr. J. U. Chukudebelu (if I were to have a second dad, it would be him)
for their patience and constructive criticism towards me and my work.
I also acknowledge my family for their unalloyed and unreserved support towards my education. I
want to specifically observe my mum, Mrs. N. N. Ibekwe (the best mum in the world) and Engr.
Nelson Uzonwa and his family for investing in me. Your efforts have finally paid off.
To my course mates, Chimaroke, Chisom and Robert. You guys are the best.
I also want to appreciate Mrs. Ijeomakanu. She surprised me sometime.
To all my friends, Cyril, Iyke Allah, Victor (76), Salata and Chinenye. I love you all!
5
DEDICATION
This work is dedicated to all those who believe in me, most especially, my mum and sisters.
And also to my future kids!
6
ABSTRACT
In the summer of 1921, a small team of physicists and geologists (William P. Haseman, J. Clarence
Karcher, Irving Perrine, and Daniel W. Ohern) performed a historical experiment near the Vines
Branch area in south-central Oklahoma using a dynamite charge as a seismic source and a special
instrument called a seismograph. The team recorded seismic waves that had traveled through the
subsurface of the earth. Analysis of the recorded data revealed seismic reflections from a boundary
between two underground rock layers.
Further analysis of the data produced an image of the subsurface, called a seismic reflection profile
that agreed with a known geologic feature. That result is widely regarded as the first proof that an
accurate image of the earth's subsurface could be made using reflected seismic waves.
Since that day till date, more technological and developmental advancements have been recorded in
the geophysical sphere the world over. Today, reflection seismology is a thriving business with the
result of the 1921 experiment being nothing compared to present day survey results.
This review attempts to take a critical look at some seismic reflection operations that have been
performed around the world, analyze the approach employed in the acquisition, processing and
interpretation of seismic data and to suggest or recommend suitable and more appropriate approaches
to specific survey objectives.
This review work begins by introducing the basic concepts and principles upon which seismic
reflection work operates, and then will take a look at seismic data acquisition methods and the basic
processes applied in transforming a noise-filled data to an output with clearer features (data
processing).
Ten (10) seismic reflection cases will be reviewed. The review will cover their
7
acquisition, processing and interpretation procedures and their results discussed. This project will
conclude by proffering recommendations as to what method or sequence a seismic crew could adopt
to address a desired result.
8
TABLE OF CONTENT
TITLE PAGE
i
CERTIFICATION
ii
ACKNOWLEDGEMENT
iii
DEDICATION
iv
ABSTRACT
v
TABLE OF CONTENT
vii
LIST OF FIGURES
xi
LIST OF TABLES
xii
CHAPTER ONE: GENERAL INTRODUCTION
1.1
Introduction
1
1.2
Elasticity Theory
2
1.3
Stress
2
1.4
Strain
4
1.5
Seismic Wave Equation
6
1.6
Seismic Waves
9
1.6.1
Body Waves
10
1.6.2
Surface Waves
11
1.7
Seismic Wave Propagation
11
1.7.1
Huygen’s Principle
12
1.7.2
Fermat’s Principle
12
1.8
Wave Attenuation
12
1.9
Seismic Velocity
13
1.10
Factors that Affect Seismic Velocity
14
1.11
Purpose of Study
14
CHAPTER TWO: SEISMIC DATA ACQUISITION AND PROCESSING
9
2.1
Introduction
15
2.2
Seismic Data Acquisition
15
2.3
Seismic Data Processing
16
2.4
Pre-Processing
17
2.4.1.
Demultiplexing
17
2.4.2
Trace Editing and Muting
19
2.4.3
Gain Recovery
2.5.
Main Processing
21
2.5.1.
Deconvolution
22
2.5.2.
CMP Sorting
25
2.5.3.
Velocity Analysis
26
2.5.4.
NMO Correction
29
2.5.5.
Dip move-out (DMO) Correction
30
2.5.6.
Statics Corrections
32
2.5.7
Residual Statics Correction (RSC)
34
2.5.8.
Stacking
35
2.5.9.
Digital Filtering
35
2.5.10.
Migration
36
2.6.
Mathematical Operations in Seismic Data Processing
40
2.6.1.
The Fourier Series/Transform
40
2.6.2.
Convolution
44
2.6.3.
Correlation
46
2.6.4.
The z-transform
47
19
CHAPTER THREE: LITERATURE REVIEW
3.1.
Introduction
3.2.
Case 1: Seismic Reflection Studies in the Skellefte
49
10
Ore District, Northern Sweden
3.3.
3.4.
3.5.
3.6.
3.7.
3.8.
3.9.
3.10.
3.11
50
Case 2: Seismic Reflection Tomography: A Case
Study of a Shallow Lake in Lake Balaton
53
Case 3: COCORP’S Deep Seismic Reflection Profiling in
the Williston Basin (Comparison of Explosive and
Vibroseis Source Energy Penetration)
57
Case 4: Reflection Seismic and Ground-Penetrating
Radar Study of -Previously Mined (Lead/Zinc) Ground,
Joplin, Missouri
61
Case 5: Seismic Processing and Velocity Assessments
in the Oil and Gas Resource Potential of the 1002 AREA,
Arctic National Wild Life Refuge (ANWR), Alaska
64
Case 6: Seismic Reflection Investigation of C-1 to F-1
Sector of the Superconducting Super Collider (SSC)
Ring West of Waxahachie, Texas
65
Case 7: Analysis of Existing Seismic Reflection Data in
South Florida for Aquifer Storage and Recovery (ASR)
Regional Study
71
Case 8: 4D Acquisition Using Two Boats in the Enfield
Oil Field Offshore, Western Australia
72
Case 9: 3D Interpretation of The Espoir Field Area,
Offshore Ivory Coast
76
Case 10: Structural and Stratigraphic Mapping of
Emi field, offshore Niger Delta, Nigeria.
78
CHAPTER FOUR: RESULT DISCUSSION
4.1.
Results of Case 1
82
4.2.
Results for Case 2
84
4.3.
Results of Case 3
86
4.4.
Results of Case 4
89
11
4.5.
Results of Case 5
90
4.6.
Results for Case 6
90
4.7.
Results for Case 7
91
4.8.
Results for Case 8
92
4.9.
Results for Case 9
92
4.10.
Results for Case 10
93
CHAPTER FIVE: DISCUSSIONS AND CONCLUSIONS
5.1.
Discussions and Conclusions
97
5.3.
Recommendation
99
REFERENCES
101
12
LIST OF FIGURES
Fig. 1.1: Components of stress acting on faces perpendicular to the x-axis.
3
Fig 2.1: Conventional seismic data processing flowchart
18
Fig 2.2 A stacking chart showing traces and common gathers
26
Fig 2.3: A CMP gather (a) containing a single event with a move-out velocity of
2264m/s; (b) shows NMO correction of (a) using appropriate velocity
(2264m/s); (c) shows an over-corrected event because a very low velocity
was used (2000m/s) and (d) shows under-correction resulting from too
high a velocity (2500m/s).
31
Fig 2.4: Illustration of elevation statics
33
Fig 2.5: Design and application of zero-phase filter in the frequency domain
38
Fig 2.6: Design and application of zero-phase filter in the time domain
38
Fig 3.1:Bedrock geological map of the Kristineberg area, with the locations
of the seismic profiles, the CDP lines and the temporary seismic stations
51
Fig 3.2: Location map of the broader study area of Lake Balaton
54
Fig 3.3. A map showing the survey area of the Williston basin and Trans-Hudson
Orogen of North Dakota
58
Fig 3.4: Map of the survey area in Joplin, Missouri
62
Fig 3.5: Location of seismic line segment C1 – F1
68
Fig 3.6: Location map of the Enfield 4D monitor survey
74
Fig 3.7: Location map of 1981-82 Ivory Coast 3-D seismic survey showing
structure of Albian unconformity
77
Fig 3.8: Map of Niger Delta showing the location of the study area (Emi field)
79
Fig 4.1: A brute stack showing (a) before and (b) after applying refraction and
residual statics, and a time and space-variant band-pass filter
83
Fig 4.2: Line drawing of the stacked section along Profile 5 (a) before and
13
(b) after migration and depth conversions
85
Fig 4.3: Raw zero offset section (512 traces with 25cm trace spacing)
containing three reflectors acquired in Balaton Lake
86
Fig 4.4: The same data in figure 4.3 after (a) band-pass filtering,
(b) gain recovery function, (c) deconvolution (optimum Wiener
deconvolution) and (d) phase-shift migration of zero-offset data
87
Fig 4.5: Reconstructed reflectors after inverting reflection times and
appropriate regularization factors have been applied. Dashed lines
represent initial reflectors and thick continuous lines represent the
final inverted geometry
88
Fig 4.6: A stacked section of data obtained from lines (a) MT10(vibroseis)
and (b) MT12 (explosive) imaging the same unmigrated sub-surface
features through 30s TWT. (c) is still MT12 imaging the upper mantle
features
89
Fig 4.7: In-line 525, crossing A-1X well location, Espoir field, and showing
clear definition of rotated fault blocks beneath Albian unconformity
94
Fig 4.8: Comparison of 2-D migrated and 3-D migrated sections across structure
drilled by well A-4X showing improved definition of erosional high on
Albian unconformity and fluid contact (flat spot)
96
LIST OF TABLE
Table 2.1: Basic Fourier transform theorems
43
Table 3.1 Data acquisition parameters for profiles 1 and 5
55
Table 3.2: A summary of the processing operations on profiles 1 and 5
56
Table 3.3: Acquisition parameters for the Lake Balaton survey
57
Table 3.4: A table showing recording parameters for both dynamite and vibroseis
Data of the 1002 area
66
Table 3.5: Acquisition parameters for the ASR studies seismic survey
73
Table 3.5. A summary of the processing steps applied to the ASR studies data
73
14
CHAPTER ONE
GENERAL INTRODUCTION
1.1.
Introduction
Seismic reflection method is the most commonly used geophysical technique which applies the
principles of seismology for stratigraphic and structural mapping beneath the ground surface to depth
of several hundred meters.
Seismology is a wide scope in geophysical science that seeks to analyze the nature of the sub-surface.
It studies movements in the sub-surface and its resultant effect on the earth’s surface. Seismological
studies are based on the travel-time and amplitudes of waves through a medium. The time it takes a
wave to travel through a medium suggests the nature and characteristics of that medium.
The origin of seismology is traced back to 132AD in the ancient Chinese kingdom when a Chinese
scientist, Chang Heng invented the first functional seismoscope (Lowrie, 2007). Before this time,
disturbances within the earth such as earthquakes and volcanoes were thought to be the wrath of
angry gods and such theories were generally accepted.
The development of seismology was somewhat slow from 1638 when Galileo described the response
of a beam to loading, to Hooke’s establishment of the law of spring in 1660. It took another 150 years
for Navier to put down the elasticity equations. Time also slipped before Cauchy and Poisson
established the modern elastic theory which formed the basis of the study of the nature of the earth.
Not until 1892 did seismic knowledge receive a rapid boost when John Milne invented a sensitive and
reliable seismograph (Lowrie, 2007).
15
Though not comparable to modern equipment, it allowed quantitative and accurate description of
earthquakes which aided the study of the earth’s seismicity and internal structure.
1.2. Elasticity theory
This theory is a very expedient prerequisite for the study of the physical principles upon which
seismic wave propagation characteristics is based. Such characteristics range from the generation of
such seismic waves, their transmission, absorption and attenuation as they travel through the earth.
Others include their reflection, refraction and diffraction characteristics at boundary surfaces. This
theory was postulated by Hooke in which he stated that ‘within the elastic limit of an elastic material,
the stress in that body is proportional to the strain’. This is Hooke’s law and forms the basis of elastic
theory which the earth materials are thought to obey. Hooke’s law is given by;
Two basic concepts are used to describe wave propagation through the earth. They are stress and
strain. Hooke’s law applies for small strains well below the elastic limit. When several stresses are
involved, each produces strain independent of the others. Hence, total strain is the sum of strains
produced by individual stresses i.e. each stress is a linear function of all stresses and vice versa.
1.3. Stress
Stress, p, is defined as force per unit area acting on a surface and is measured in
. If the force F
exceeds a value called the proportionality limit, Hooke’s law no longer holds and the body suffers
permanent deformation and does not return to its original shape and size when the stress is removed.
If force, F is normal to the surface, we have normal stress. If F is tangential to the surface, we have
16
shearing stress. If F is neither parallel nor perpendicular to element of surface, it can be resolved into
normal or shearing stresses.
Fig. 1.1 shows a small element of volume inside a stressed body in which stresses acting on the two
faces normal to the x-axis are shown.
, denotes a stress parallel to the y-axis acting on a face
perpendicular to the x-axis.
z
x
Fig.1.1. Components of stress acting on faces perpendicular to the x-axis.
If the stresses are balanced, the medium is said to be in equilibrium, i.e.
acting on a
plane a-b-c-d must be equal and opposite to the corresponding stresses shown on the opposite face
e-f-g-h. For a body in equilibrium, moment = 0.
The force, F is defined by its components in
directions by
of the surface element is characterized by its area A, with
to the x-axis. The component
by
, and also shear stresses
being the small surface with area normal
acting normal to the surface
and
respectively. The size
produces a normal stress, denoted
in the y and z-axis. That is;
17
Same applies also for components of
resulting in the (3×3) matrix
The stress matrix shows nine stress components which completely define the state of stress of a body.
If the forces are balanced, it will contain only six independent elements.
1.4. Strain
A change in shape and dimension of a body as a result of the effect of stress is termed strain. Strains
are of different fundamental types, normal strain and shearing strain. In addition to these strains, the
body is subjected to simple rotation about its three axes.
Longitudinal strain is defined as the fractional change in length of an element along a specific
direction. The component of strain parallel to the x-axis resulting from a small displacement parallel
to the x-axis is denoted by
The longitudinal strains
where
and is given by Dobrin and Savit (1988) as;
are of opposite signs but proportional to
is the constant of proportionality called Poisson’s ratio and is given by;
. i.e
18
Dilatation,
which is defined as the fractional change in volume per unit volume of an element in the
limit when its surface area decreases to zero and is given (Lowrie, 2007) by;
Shear strain is a products of the shear components of stress (
) due to changes in the
angular relationship between parts of a body. When for instance, a square is sheared parallel to the xaxis, the side parallel to the y-axis rotates through an angle
axis, the side parallel to the x-axis rotates through an angle
. When it is sheared parallel to the y.
The shear strain in the x-y plane is half the total angular distortion (Lowrie, 2007). i.e.
It could be illustrated in a symmetric 3×3 strain matrix as
19
The deformation of a body gives rise to a relationship between stress and strain. The ratio of such
relationship defines the elastic modulus of that body. The unit for elastic moduli is
(Newton
per meter squared). Elastic moduli is defined for three types of deformation;
i.
Young’s modulus, E: This results from deformation due to extension. i.e.
,
where E is Young’s modulus and is the constant of proportionality.
ii.
Shear or rigidity modulus, : This results from shear deformation. i.e.
,
where constant of proportionality is shear modulus .
In terms of Young’s modulus and Poisson’s ratio,
iii.
is given by
Bulk Modulus, K : This is the ratio of hydrostatic pressure to the dilatation. That is
where P is the hydrostatic pressure, K is Bulk modulus and
In terms of Young’s modulus and Poisson’s ratio,
is the dilatation.
is given by Telford (1990) as,
(1.8)
1.5. Seismic wave equation
To derive the wave equation for both P and S waves, we consider the stress-strain diagram (fig. 1.1),
from which we obtain the stresses on the front face e-f-g-h, as,
(1.9)
Because these stresses are opposite to those acting on the rear face a – b – c - d, some stresses balance
and cancel out. Unbalanced stresses are given as;
20
Hence, the net forces acting per unit volume in the x, y, z directions have values;
Other faces hold similar expressions. Therefore, we derive total force per unit volume in the xdirection as
We apply Newton’s second law to the total force per unit volume in the x-direction to set;
(1.10)
where U is displacement in x-direction and
is the body density.
Equation 1.10 relates the displacement to the stresses.
But by Hooke’s law,
.
Likewise,
We substitute these expressions in equation (1.11) also noting that
,
where
(1.12)
is called dilatation and is defined as the fractional change in volume per unit volume of an
element in the limit when its surface area decreases to zero.
We therefore obtain the expression;
21
Similarly, the corresponding expressions for
and
are
By multiplying (1.13 a) by i, (1.13 b) by j and (1.13 c) by k, we obtain
By defining displacement as
becomes
Equation (1.15) is the equation of wave propagation in a homogeneous isotropic elastic solid. It is an
approximate expression which neglects body forces such as gravity and velocity gradient terms and
assumes a linear isotropic earth model.
If we take the divergence of both sides of (1.15), we obtain
But,
Equation (1.16) then becomes,
=
Equation (1.17) is the P-wave equation in which dilatation propagates with a velocity,
22
.
For S-wave equation, we take the curl of both sides of (1.15)
But the rotational vector,
=
is the curl of displacement vector. i.e.
Substitute (1.19) into (1.18) noting that the curl of a gradient is identically zero, i.e.
to
obtain;
Or
Equation (1.21) is the S-wave equation from which S-wave velocity is given by
.
1.6. Seismic waves
Seismic waves are disturbances that propagate through a medium. When these waves penetrate a
medium, the particles in that medium are displaced and deformed elastically to permit the passage of
the waves. There are two types of seismic waves and they are surface waves and body waves.
23
1.6.1. Body waves
These are waves or energy that propagates through the body of the medium. They gradually transform
from spherical waves to plane waves as the distance from its source increases due to the reduction in
curvature of its wave front. Body waves are of two types, longitudinal and transverse waves.
Longitudinal or compressional waves are waves that pass through a medium as a series of rarefactions (dilatations) and compressions.
Transverse or shear waves are concerned with the vibrations along the y and z directions which are
parallel to the wavefront and transverse to the direction of propagation in the x-direction.
The general shear wave motion within the plane of the wavefront can be resolved into two orthogonal
components, the vertical and horizontal components. For the vertical component in the z-direction,
the equation;
describes the shear wave with particles displaced in the z-direction. This wave is considered to be
polarized in the vertical plane. The horizontal component in the y-axis is analogous to the above
equation with particles displaced in the y-direction. The resultant wave is horizontally polarized and
can be called SH-wave. Shear wave in the x-direction is given by the equation (Lowrie, 2007),
where ψ is the rotational vector and
.
24
Longitudinal waves are the fastest of all seismic waves and so arrive first at a detector (seismometer)
when an earthquake occurs. The first arrivals are called primary waves (P-waves). P-waves travel
through solid, liquid and gas because they are compressible. Transverse waves on the other hand are
much slower and so constitute later arrivals at recording stations. These waves are called secondary
waves (S-waves). They penetrate solids but can’t transmit through liquids and gases.
1.6.2. Surface waves
These are disturbances propagating close to a free surface of a medium. Surface waves are classified
into several types, the most common types are Love waves and Rayleigh waves.
Love wave are waves that can propagate only in a velocity layered medium. The particle motion is
horizontal and perpendicular to the direction of propagation, with its amplitude decreasing with depth.
For a two-layer medium with S-wave velocities
closer to the slower velocity
and
, with Love waves of short wavelengths
and long wavelengths travelling at speed closer to the faster velocity,
. Love waves are dispersive (Lowrie, 2007).
At the surface of homogenous and isotropic half space, the particles in Rayleigh waves are polarized
to vibrate in the vertical plane containing the direction of wave motion. The resulting particle motion
is a combination of the P and SV-vibrations. Rayleigh waves are non-dispersive. Raleigh waves
(often called ground roll) are a main source of signal generated noise on seismic reflection records.
1.7. Seismic wave propagation
Propagation of waves through a medium is as a result of the periodic elastic displacements of the
particles in that medium and its progress is determined by the advancement of the wavefront. Two
principles can be used to handle this. These principles are Huygen’s principle and Fermat’s principle.
25
1.7.1. Huygen’s Principle:
This principle as proposed by a Dutch mathematician/physicist, Christiaan Huygens in 1678 describes
the behavior of wavefronts. It is based on simple geometrical construction. It states that “all points on
a wavefront can be regarded as source points for the production of new spherical waves. The new
wavefront is the tangential surface or envelope of the secondary wavelets”. This method permits the
calculation of the future position of a wavefront if the present position is known. This principle is
applied in reflection and refraction, mainly at their boundaries.
1.7.2. Fermat’s Principle:
Formulated by a French mathematician, Pierre de Fermat, this principle states that of the many
possible paths between two points, the seismic ray will follow the path that gives the shortest traveltime between the points. This principle describes the geometry of the ray paths at an interface.
1.8. Wave attenuation
This is the decrease of amplitude of a wave with increasing distance from the source. This reduction
is due to either the geometry of the propagation of the wave or the anelastic property of the medium
through which the wave travel. Amplitude attenuation is more rapid in body waves than in surface
waves. They have a corresponding attenuation of
respectively.
Absorption of energy due to imperfect elastic property of the medium causes attenuation. This is as a
result of energy loss amongst rock particles and can also be called anelastic damping. Damping of
waves is described by a parameter called quality factor which is defined as the fractional loss of
energy per cycle. Damped amplitude of a seismic wave at distance r from its source is given (Lowrie,
2007) by;
26
where D is the distance within which the amplitude falls to
of its original value.
Its follows from equation (1.24) that absorption co-efficient is inversely proportional to the
wavelength
and is dependent on the frequency of the signal.
1.9. Seismic velocity
Velocity plays an important role in data interpretation. The knowledge of seismic velocity values is
very necessary in the determination of depth and dip of reflectors and refractors and in the general
assessment of the nature of the rocks and interstitial fluids in the sub-surface. Velocity is obtained
from measurements taken in the field, well logging or from samples in the laboratory.
There are different types of velocity. These include average velocity, root mean square velocity,
interval velocity, phase velocity, stacking velocity, instantaneous velocity and apparent velocity
Average velocity is defined as the distance travelled divided by the time taken to traverse path. Root
mean square velocity refers to the velocity of a specific ray path as it travels through layers. Interval
velocity on the other hand is the average velocity over some interval of the travel path. Phase velocity
refers to the velocity of perpendicular distances between phases of a wavefront. Stacking velocity is
the velocity value determined by velocity analysis usually made by finding the best fitting hyperbola
to data that are not perfectly hyperbolic. It is used to correct the quasi-hyperbolic primary reflection
event to time alignment with zero offset. Instantaneous velocity is the velocity with which a
wavefront passes through a point measured in the direction of travel. Apparent velocity refers to the
apparent speed of a given phase in a particular direction usually, spread direction on the surface.
27
1.10. Factors that affect seismic velocity
Due to the non-homogenous nature of the earth, seismic velocity varies from location to location and
an understanding of these factors helps us foresee the kinds of velocity variations to expect in such
areas. Such factors include effect of lithology (that is rock type), effect of density (where velocity is
higher in more dense rocks), porosity of rocks, depth of burial (in which case the greater the depth of
burial, the more the pressure and consequently, higher velocity). The age of the rock type is another
factor. Others are effect of interstitial fluid where more pore fluid depicts lower velocity, temperature,
frequency etc.
1.11. Purpose of study
The purpose of this study is to intensively review the different approaches applied to seismic
reflection survey and the results obtained during several operations as carried out in different parts of
the world. The study and comparison of these seismic reflection operations and the results obtained is
aimed at suggesting a more appropriate procedure in seismic data collection, processing and
interpretation, for yielding more robust results.
28
CHAPTER TWO
SEISMIC DATA ACQUISITION AND PROCESSING
2.1. Introduction
Over the decade, seismology has become the key tool to exploration and developmental successes.
With advancements in computer technology, data processing has increasingly acquired a competitive
edge and resulted in the production of clearer and more detailed seismic sections. There is no specific
or generally accepted mode of processing but the choice of a processing method depends on the
information hoped to be obtained or simply the choice of the analyst. The choice of equipment also
determines processing sequence. We should also bear in mind that there are also a variety of
objectives for which a data can be used. Data processing to a certain degree can be adjusted to meet
specific requirement. Seismic data processing is characterized by a sequence of steps where for each
of these steps, there exists a multitude of different approaches.
2.2. Seismic data acquisition
Seismic data acquisition requires a controlled source of energy or signal, usually an explosion or
vibration sent into the ground at a known time and location. By taking note of the time it takes the
signal to be reflected back from the boundaries between rocks to a receiver on the earth’s surface
(usually a geophone for land use and a hydrophone for marine), the depths and features of the subsurface interfaces can be estimated after some processing and analyses.
Seismic data could be acquired on land (land survey) or over an aquatic environment (marine
survey). Both methods employ the same principle of sending signals into the ground and recording
the two-way travel time even though the equipment used differs. Energy sources for land survey
29
include explosives, impact sources (weight drop), vibrators and even hammer for shallow smallscale exploration while for marine surveys, airgun, watergun, marine vibrators, aquapulse,
maxipulse, flexotir etc. are employed.
Several field procedures are in use, each being distinguished by the different layout of the
geophones relative to the shot-point. Continuous profiling is usually the routine in which the shotpoint and geophones are moved a predetermined distance after each shot is taken. Continuous
profiling could be conventional i.e. when a reflecting point is sampled once, or redundant, when it is
sampled more than once. Continuous profiling aims at reducing seismic noise. Split spreading is the
most common form of convectional coverage in which the geophones are symmetrically spread on
either side of the shot-point while the common mid-point (CMP) method is an illustration of the
redundant coverage in which shots are fired at different points and each shot has a sub-surface
coverage such that by repeatedly moving the shot-point and geophone arrays, each reflecting point
on the interface is sampled severally
As the geophones/hydrophones detect the arrival of seismic waves, the signals are converted to a
digital form in the present day, unlike in the early systems which recorded analogue signals directly to
magnetic tapes or photographic films. The recorded signals are then displayed by a computer as
seismograms for processing and analysis.
2.3. Seismic data processing
The main goal of seismic processing is to construct high quality reflectivity images of the investigated
sub-surface. Bearing in mind the inhomogenity of the earth, several factors such as environmental and
demographic restrictions can have significant impact on data quality. Other factors such as field
geometry and conditions of recording could also pull strings. Having knowledge of these restricting
factors and the need to have a clearer picture of the sub-surface to compensate for the whole effort
30
and finance put in is necessary. Therefore, obtained data undergo several stages of processing, all
primarily aimed at increasing the signal to noise ratio, thus, producing genuine reflections which
could be interpreted.
Seismic data processing comprises of three main stages; the pre-processing stage, the main processing
stage and the final processing stage as illustrated in fig 2.1.
2.4. Pre-processing
Pre-processing, as can be coined from its name, are processing operations that are carried out on the
field data before the main processing. Such operations include demultiplexing (reformatting), muting,
gain application, filtering and identing. Let us consider them briefly.
2.4.1. Demultiplexing
Field data are recorded in multiplexed form. A multiplexed data is a continuous stream of seismic
samples for the whole seismic record. i.e. in rows of samples at the same time at consecutive
channels. A digital sample is recorded for each seismic trace at a pre-selected sampling rate (2ms or
4ms) in rows of samples at several channels, consecutively. Demultiplexing sorts these data into
columns of samples, i.e. all the time samples in one channel, followed by those in the next channel
and so on.
31
Fig. 2.1. Conventional seismic data processing flowchart (Dobrin and Savit 1988).
32
It can be said that demultiplexing is the separating of all the samples to produce a time sequence for
each geophone. Mathematically, it could be seen as transposing a big matrix such that the column of
the resulting matrix can be read as seismic traces recorded at different offsets with a common shotpoint (Hatton et al, 1986).
Multiplexed data are time sequential while demultiplexed data are trace sequential.
2.4.2. Trace editing and muting
This is a manual cleaning up process of the data. Dead traces, noisy traces (ground rolls, direct
arrivals), mono-frequency signals with transient glitches and less relevant events are deleted and
polarity of reverse traces switched. Frequency distorted zones are also muted to avoid them
suppressing shallow events.
2.4.3. Gain recovery
This function is applied to data to correct for the amplitude effect of spherical wavefront divergence.
This involves the application of a geometric spreading function which depends on travel-time and an
average primary velocity function which is associated with primary reflections in the survey area. It
will also require an exponential gain function to compensate for attenuation losses.
Gain is a time variant scaling in which the scaling function is based on desired criterion. The gain
function for geometric spread compensation is given by Yilmaz, (1987) as;
(2.1)
33
Applying the geometric spread correction amplifies frequencies at the late times of a spectrum but
does not restore the amplitudes of high frequencies as much as it does for low frequencies. To correct
for their amplitudes, we apply deconvolution or time-variant spectral whitening. Gain application to
seismic data also enhances display quality.
There are three common types of gain: the programmed gain control (PGC), the RMS amplitude
automatic gain control and the instantaneous automatic gain control (Hatton et al., 1986).
The programmed gain control (PGC) is the simplest form of gain. The PGC function is applied to all
traces in a gather to preserve the relative amplitude variations in the lateral direction. A gain function
can be defined by interpolating between some scalar values specified at specific time samples.
The RMS amplitude AGC has its gain function based on rms amplitude on an input trace within a
specified time gate. To obtain this, first the amplitude of each sample in a gate is squared, then its
mean value computed and then its square root taken. Usually, the value of the gain function at the
centre of the gate is the ratio of the desired rms amplitude to the central rms value.
Instantaneous AGC is one of the most common types used. To compensate its gain function, its mean
absolute value of trace amplitudes is first computed with a specified time gate, and then the ratio of
the desired rms level to this mean value is taken to be its gain function. The gate is moved one sample
down the trace to compute the gain functions of subsequent time samples. In this case, no
interpolation is required.
This ends the procedures performed in the pre-processing stage. At this stage, a large percent of noise
in the signal must have been eliminated, parameters properly idented and a fairly better quality of
display achieved due to applied gain. The data is now ready for the main processing.
34
2.5. Main processing
There are three basic stages in seismic data processing. These stages are deconvolution, stacking and
migration. Each of these satisfies the goal of data processing which is improvement of the temporal
resolution of seismic data (deconvolution), enhanced signal to noise ratio (stacking) and improvement
of lateral resolution (migration).
Other processes involved in the main processing stage include normal move-out (NMO) correction,
elevation correction, dip move-out (DMO) correction and frequency filtering.
NMO correction compensates for differences in travel times due to differences in source-receiver
offset so that all traces having the same common mid-point (CMP) would appear to have zero offsets
and combined as one. This combination of traces having the same CMP is referred to as stacking.
CMPs are sorted out using a binning grid superimposed over the survey area. The grid is usually
oriented along the actual shot and geophone lines or for marine prospect, along the shooting direction.
The dimensions of the bins are usually based on the shot spacing and geophone spacing. A number of
traces or CMP will fall within each grid cell or bin. These traces are assumed to be imaging the same
general area of the sub-surface.
After applying NMO and static corrections, there is need to perform the dip move-out (DMO)
correction to correct for travel time differences due to dipping reflectors. The traces are then stacked
within each bin to produce a single trace usually at the bin centroid. After stacking, other multi-trace
processes such as migration are possible.
Some factors can just not be left out. In the processing stage, about the most important element is the
velocity information. Without this information, NMO, DMO and static corrections would not be
feasible. CMP stacking, migration and time-depth conversion will also not be possible. This makes it
very expedient.
35
Velocity information also helps in rock identification. Before the computer era, velocity
determination used to be analytic but these days, velocity can be determined directly from reflection
data, well logs, sonic logs etc.
Filtering is another sequence in the processing stage. As the name implies, it simply means
eliminating unwanted data. But in this case, it is usually data outside a particular range. There are
several types of filters and they will be discussed in a few pages.
2.5.1. Deconvolution
Deconvolution aims at generating a smoothly tampered wavelet with a wide spectrum. It improves
temporal resolution by compressing effective source wavelet contained in the seismic trace to a spike
of zero-lag. This is called spiking deconvolution. Since this broadens the spectrum of seismic data,
traces tend to contain much more high frequency energy after deconvolution. To have a better
understanding of the deconvolution process, let us take a brief look at convolution and its model.
The convolution model attempts to explain how seismic trace is formed. It approximates the earth by
a linear system. A linear system is one whose output, say
with its response
, such that
In the convolution method, the seismic trace
where
is given by the convolution of its input
is the source wavelet and
is given by
is the earth’s response.
36
If the noise component
is present, equation (2.3) becomes
Convolution is a time-domain operation which involves the replacement of each element on an input
function with an output function and superimposing the output on the input function. For linear
systems, which the convolution model approximates the earth to, the output is time invariant and
directly proportional to the input. We can convert from output to input if the impulse response of the
system is known.
The forward convolution model of the seismic trace is used to compute the synthetic seismogram
given the source wavelet and earth’s reflectivity. The convolution model of the seismic trace is widely
accepted because it agrees well with the observed seismic traces even though it assumes a lot.
Deconvolution, which is the inverse convolution model, is used to compute the earth’s reflectivity
given the seismic trace,
If the output
. Seismic trace here refers to the output.
is known and we want to regain our initial input signal as it was before it was
modified by the filter, we need to find another filter called the inverse filter with a response, say
through which
would pass to recover
. This process of cancelling the effect of a filter with
another filter is known as deconvolution or inverse filtering.
In the time domain, deconvolution is given by;
,
37
In the frequency domain, it is given by;
Deconvolution is divided into spiking deconvolution and predictive deconvolution. If the source
wavelet is known, then the deconvolution is deterministic and we use spiking deconvolution (inverse
filtering) to find the earth’s response. If source wavelet is unknown, then it is said to be statistical and
in such cases, predictive deconvolution is applied.
Spiking deconvolution is mathematically accomplished using the z-transform. Predictive
deconvolution on the other hand cannot be solved using normal equations except we require the
desired output to be a time-advanced version
of the input
, where
is the prediction lag
which is known (Telford et al., 1990). Its prediction lag gap is equal to the first or second zero
crossing of the autocorrelation function. As its name implies, predictive deconvolution is used to
predict the value of an input at some future time. It also gives room to calculate the error series.
Predictive deconvolution is also a tool for marine multiple suppression.
Deconvolution is a filtering process designed on each trace even before NMO correction and stacking
so that distortions due to time variant operations in the processing sequence are eliminated. A
deconvolution operator aims at summing all the traces so that they represent the same propagating
wavelet.
Deconvolution techniques used in conventional processing are based on the optimum Wiener filtering
theory and all filtering algorithms based on this theory are known as Wiener-Levinson algorithms
(Robinson and Coruh,1988).
38
2.5.2. CMP sorting
This is the transformation of data from shot-receiver to mid-point-offset coordinates. It requires
information about the field geometry. Conventional seismic data processing is done using midpointoffset coordinates and this is achieved by first sorting the data into CMP gathers. Traces with the
same mid-points are grouped together to form a CMP gather. Fig. 2.2 shows a stacking chart which
will make our understanding of these gathers clearer.
Stacking charts are useful when setting up the geometry of a line for pre-processing and makes
identification of missing traces easy to find.
Due to the large number of traces involved in CMP acquisition, a stacking chart is used to keep track
of values. A Stacking chart could either be surface, which has geophone location g, as one coordinate
or source location s, as another or sub-surface, which has the trace plotted at
(Telford
et al., 1990).
Stacking charts are useful in making static and NMO corrections and ensuring proper trace tracking.
Displays of the data in different directions are called gathers or domains. We have the common
source gather, common geophone gather, common mid-point gather and common offset gather. All
these prove useful in the study of noise such as multiples, ghosts, converted waves etc. For horizontal
interfaces, a source placed at the centre of a geophone spread will give a symmetrical curve about the
source point. If however the reflector has a uniform dip, the path of reflected ray on the down-dip side
of the shot-point will be longer than those on the up-dip side and this has corresponding effect on the
travel times. If reflected records are not corrected for layer dip, an error results in plotting the position
of dipping beds.
39
The CMP technique uses redundant recording to improve signal amplitude to noise ratio by a
theoretical factor of
, where N is the total number of elements in the recording system. i.e.
(multiples per trace) assuming identical CMP gather and that the noise of each trace is
uncorrelated. This technique also attenuates coherent noise. This is possible due to the difference in
stacking velocities of coherent noise and reflected signals.
Fig.2.2. A stacking chart showing traces and common gathers (Telford et al., 1990).
2.5.3. Velocity analysis
The objective of velocity analysis is to determine the seismic velocity of layers in the sub-surface.
Seismic velocities are used in many processing and interpretation stages such as mentioned in the
introduction of this chapter. Velocity analysis is performed on selected gathers or group of gathers.
40
The output of one type of velocity analysis is a table of numbers as a function of velocity versus twoway zero offset time (velocity spectrum). These numbers represent some measure of coherency along
the hyperbolic trajectories governed by velocity, offset and travel-time. Velocity time pairs are
selected from velocity spectrum based on maximum coherency peaks. These velocity functions are
then spatially interpolated between the analysis points across the entire profile. When velocity picks
become inaccurate due to complex structures, the data could be stacked with a range of constant
velocities which are then used to pick the velocities.
NMO and stacking velocities can be derived from T-X data. The T-X curve of a single constant
velocity horizontal layer is a perfect hyperbola given by
where
is the two-way travel time at offset
and
is the two-way travel time at zero offset,
is
the layer velocity. (Hatton et al., 1986).
The rms velocity can be defined in terms of the true T-X curve as the square root of the reciprocal of
the coefficient of the
term in the series approximation of the exact
layers. It can be shown that
tangent to the exact
where c is the wave velocity
curve for multiple
is equal to the square root of the reciprocal of the slope of the
curve at x = 0, that is;
41
The stacking velocity,
is found from the T-X data by fitting a best fit hyperbola to the true (non-
hyperbolic) T-X curve, which takes the form;
When fitting a hyperbola to a true T-X curve, the constant term should be equal to
, a condition
that must be satisfied.
NMO velocity:
is found from the T-X data by searching for the velocity that will best NMO-
correct a certain reflection.
can be related to sub-surface properties but
cannot.
However, at small offsets,
are approximately equal to
.
We can therefore relate
to layer properties directly using the Dix formula (Constain,
2004);
where
are bottom and top layers and
are zero offset travel-times at
the top and bottom of Nth layer.
We can also find velocity by producing panels of stacked traces (50 or more) with constant velocity.
This method is called constant velocity stacks (CVS). In this method, a selected CMP gather is
repeatedly NMO-corrected using a range of constant velocity values. The NMO corrected values are
displayed side by side and the velocity that best flattens the event is chosen as the
The CVS method is especially useful in areas with complex structures.
of that event.
42
The velocity spectrum attempts to find the stacking velocity to each reflector. It maps the T-X data of
a single CMP gather onto the velocity-spectrum plane where the vertical axis is the
horizontal axis is
and the
. The velocity spectrum method is more suitable for noise contaminated datasets.
Spread length, stacking fold, S/N ratio, time gate length etc. are some factors that limit the accuracy
and resolution of this method.
2.5.4. NMO correction
NMO is a non-linear stretching of the seismic time axis. To remove travel-time components due to
source-receiver offset, NMO correction is applied. NMO correction prepares the data for stacking and
finds the NMO velocity to the reflector.
After NMO corrections, the events are noticed to be virtually flattened across the offset range. The
resultant effect of this is the stretching of traces in a time-varying manner consequently, shifting the
frequency component towards the left end of its spectrum. If a high velocity is used for NMO
correction, the event would be under-corrected or over-corrected if a lower velocity is used (fig.2.3).
The right velocity will align the event horizontally at
.
Because velocity increases with depth, the NMO correction applied for a later
that applied to an earlier
. Therefore, the two events with a time separation equal to
NMO correction will have a separation of
is NMO correction time,
before
after NMO correction. This is called NMO
stretching and it results in frequency distortion. Stretching is quantified as
where
will be smaller than
is zero offset travel time
43
It is only but wise to remove the stretching before stacking. This is done by muting the considerably
stretched zones from the gather.
In the case of a single horizontal layer with constant velocity, the T-X curve is exactly a hyperbola
given (Hatton et al., 1986) by;
We have NMO correction given (Hatton et al., 1986) by;
NMO correction is derived by subtracting
from
.
2.5.5. Dip move-out (DMO) correction
The CMP method explicitly assumes two rather oversimplifications.
1.
That the seismic velocity is constant
2.
That the sub-surface reflectors are horizontal
The latter assumption is a clear violation when processing data over dipping reflectors. In the
presence of dip, the reflections in a CMP are described by a hyperbolic curve with the apex located at
the offset from the source. The DMO correction attempts to correct the apex to zero offset which
allows for better velocity analysis and stacking.
44
Fig.2.3.A CMP gather (a) containing a single event with a move-out velocity of 2264m/s; (b) shows
NMO correction of (a) using appropriate velocity (2264m/s); (c) shows an over-corrected event
because a very low velocity was used (2000m/s) and (d) shows under-correction resulting from too
high a velocity (2500m/s). (Yilmaz, 1987)
The travel-time for reflection from a flat reflector is given by;
where T is the reflection time,
is zero-offset two-way time to reflector and h is reflector depth
(Robinson and Coruh, 1988).
For dipping reflectors,
where is the angle of dip (Lowrie, 2007).
45
Equation 2.15 can be expressed in a hyperbolic form as;
with axis of symmetry given by
.
The benefits of DMO correction include noise filtering through dip filtering, minimized reflection
point scatter, improved velocity analysis, improved ties with crossing lines etc.
2.5.6. Statics corrections
CMP gathers don’t always conform to a perfect hyperbolic trajectory. A cause of this could be as a
result of near surface irregularities. Lateral velocity variations could cause a reflection event to switch
from long offset traces to short offset traces (negative move-out) as a result of complex overburden.
Static correction is targeted towards compensating for elevation differences such that all traces are
corrected to a datum level by removing the calculated travel times from the source to the datum and
from the receiver to the datum. Fig. 2.4 illustrates this concept.
Assuming the reflection paths are vertical, travel-time of wave from source down to datum is
The correction for the travel-time for a geophone at the source point
is given by
46
where
Fig.2.4. Illustration of elevation statics (Yilmaz, 1987).
Subtracting
from the arrival time is the equivalent to placing the shot and receiver groups on a
datum plane. Elevation correction is given (Yilmaz, 1987) by;
.
Another form of static correction is the near surface (weathering) correction. It corrects for the effect
of variable thickness and lateral variation of the weathering layer. Methods used to correct for these
effects are up-hole surveys, refraction statics and residual statics. The term static is used because the
time shifts applied are constant for the entire trace.
47
2.5.7. Residual statics correction (RSC)
Of course, little random errors in static alignment on traces remain after static corrections are applied.
These little random errors are called residual statics and are obvious in corrected gather. Residual
statics cause dim spots along reflection horizon as well as false structures on stacked sections which
could be misleading. RSC tends to better align CMP gathers with travel-time deviation and usually,
velocity analysis is repeated to improve velocity picks after it is applied.
In general, RSC involves three phases: these phases include picking (calculating) the total residual
time shift
, decomposition of
into receiver, source, structural and residual terms and
application of derived source and receiver terms of travel-times on pre-NMO-corrected gathers.
RSC is basically effective in estimating short wavelength statics with the most widely used method
being the surface consistent method. This method assumes that static shifts are time delays that
depend only on the shot-receiver location on the surface and not the ray-paths in the sub-surface. The
total residual time shift
can be expressed as
. (Yilmaz,1987).
is a structural term while
is a hyperbolic term.
To minimize the error energy, we could use the least squares approach given by Yilmaz (1987) as;
48
2.5.8. Stacking
Stacking simply means summing up of individual traces to produce one stacked trace that represents
the CMP. The stacking process involves making a numerical average of all the traces samples in a
CMP gather for each sample time. By this averaging process, non-coherent events add out of phase
and are cancelled out. Stacking n traces enhances the S/N ratio by
displayed with the CMP number in the horizontal direction and
. The stacked section is
in the vertical direction. Since
stacking enhances horizontal events in a CMP gather, anything not horizontal is suppressed.
Proper measures must be observed before stacking since it would be almost impossible to undo the
process.
2.5.9. Digital filtering
To an extent, we can regard all pre-processing and the main processing stages as filtering processes
since they are directed towards eliminating unwanted signals. We can therefore describe a filter as a
system which discriminates against a certain range of its input. Frequency and apparent surface
velocity are two basic properties upon which filters base their discrimination.
In frequency filtering, a zero-phase band limited wavelet can be used to filter a seismic trace. The
filter operator is the time domain representation of the wavelet, while the filter coefficients are the
individual time samples of this operator. The above described process is termed zero-phase frequency
filtering. It does not modify the phase spectrum of the input trace but limits the band of its amplitude
spectrum.
49
Frequency domain filtering involves multiplying the amplitude spectrum of the input seismic trace by
that of the filter operator while in the time domain. It involves convolving the filter operator with the
input time series. The summary of both frequency and time domain filtering is illustrated in figures
(2.5) and (2.6). Both formulations are based on the concept that convolution in the time domain is
equivalent to multiplication in the frequency domain and vice versa (Bracewell, 1965).
Frequency filtering could be in the form of band pass which only extracts a defined band of frequency
but does not alter its phase. Examples of frequency filters are high-pass, low-pass, band-pass and
notch frequency filters. Band-pass filters are the most used.
In velocity (f-k) filtering, the f-k plane allows us to separate events that dip in the (t,x) plane by their
dips. This consequently eliminates certain types of noise from a data. This filter suppresses coherent
linear noise arising from side-scattering energy in stacked common shot gathers.
The basis for f-k filtering is that events with the same dip in the t-x plane are mapped into a single
line in the radial direction of the f-k plane irrespective of their location on the t-x plane. This results in
the defining and application of a reject fan in the transform domain and then inverse transposing the
data back to the (t,x) domain.
F-k dip filtering is recommendably applied on shot gathers instead of CMP gathers. Its application
after stacking further suppresses noise. Likewise, coherent linear noise on stacked data can be
suppressed by post-stack migration process which also incorporates dip filtering.
2.5.10. Migration
The goal of migration is to make the stacked section as close to the geological cross-section as
possible. When displayed, the stacks appear to be a geological image. Thus, migration is also called
seismic imaging.
50
Migration is an inverse wave scattering calculation that relocates seismic reflections and diffractions
to the location of their origin, arranging data laterally in the image volume. It is inherently a 2-D or 3D procedure and of course, the 3-D migration produces a more accurate seismic image. Migration
operation collapses subtle diffractions associated with growth faults and also unties ‘bowties’ into
synclines.
Migration could either be pre-stack or post-stack. Pre-stack migration is recommended for survey
areas with complex structures such that each seismic trace is migrated solely before stacking.
Velocity analysis algorithms in pre-stack migration allow processors to improve their velocity models
between migration iterations. Its major disadvantage is the large data size to be migrated. Post-stack
migration involves migrating a stacked section.
Besides distinctions between 2-D versus 3-D and pre-stack versus post-stack migrations, migration
implementations are also distinguished as either time migration or depth migration.
Time migration assumes that local variations in velocity is a function of depth alone and usually
ignores the refraction that occurs when rays cross the velocity boundaries. When lateral variations are
much, time migration is not recommended, the migration algorithms for time migration are quite
simple.
51
Fig.2.5. Design and application of zero-phase filter in the frequency domain (Yilmaz. O,. 1987).
Fig.
2.6. Design and application of zero-phase filter in the time domain (Yilmaz, O., 1987).
52
Depth migration on the other hand assumes a known velocity model and estimates the correct shape
of diffractions by ray-tracing or wave front modeling. It is a better option for areas with complex
structures.
Migrated sections are commonly displayed in time rather than depth to avoid errors introduced by
inaccurate time to depth conversion and also to facilitate the comparison of migrated section with
unmigrated section which is usually displayed in time.
In 2-D migration, we migrate the data once along the profile while for 3-D migration, we first migrate
the data in the in-line direction and then migrate the already migrated data in the cross-line direction.
This is the two-pass 3-D migration. One-pass 3-D migration can be done using a downward
continuation approach (Yilmaz, 1987).In this technique, also referred to as finite difference method,
one uses the seismic data as recorded at the surface datum plane to calculate what it would have
looked like if it were recorded on a plane one depth step deeper in the earth. The same is done for
deeper steps until one approaches the actual depth of a diffractor.
As one steps closer, the diffraction hyperbola becomes smaller and collapses eventually at a point
when t = 0, i.e. the diffractor depth. The diffractor is said to be imaged.
The above explained technique is one among other migration algorithms that have been developed to
ease degrees of complexities in the process. Others include Kirchoff migration and f-k migration.
Kirchoff migration employs Huygen’s principle (1.7.1). Here, diffraction energy is collected along
diffraction curves. Kirchoff migration assumes straight rays and takes into account amplitude and
phase corrections that can be derived from wave equations. The purpose of Kirchoff migration is to
sum up the energy produced by every Huygen’s secondary source and map it into its point of
generation.
53
F-k migration uses the 2-D Fourier transform to convert the input t-x section into an f-k section.
Migration of seismic data is carried out in the frequency domain. The seismic time section is first
converted to an apparent depth section. In the apparent depth mode, a constant velocity is assigned to
all reflections. Then, this seismic data is two-dimensionally transformed into frequency and wave
number domain. The inverse 2-D FT provides the migrated t-x section
2.6. Mathematical operations in seismic data processing
Three types of mathematical operations constitute the heart of most data processing. They are Fourier
transform, convolution and correlation, however, the z-transform can be included as another
mathematical operation.
Fourier transform (FT) is used to convert from the time domain to the frequency domain and vice
versa. It can also be used alongside other transforms to convert into and out of other domains.
Convolution operation replaces each element of an input with a scaled output function. This operation
better explains the limitations placed on sampling and signal reconstitution.
Correlation is a method of measuring similarities between two data sets by determining the time shift
that will maximize such similarities. It is also used to extract short signals of known wave shape from
long train waves as used in vibroseis processing. Auto-correlation is a situation where a data is
correlated by itself and cross-correlation occurs when different data are correlated.
2.6.1. The Fourier series/transform
Consider a mono-frequency sinusoidal function
given by
54
A time-dependent (periodic) signal
can be analyzed to an infinite series of sinusoids, each having
its own frequency, peak amplitude and phase using the Fourier series by;
The basic Fourier series expansion of a function
is given as
(2.24)
Where for any non-negative integer,
It can be expressed in exponential form as;
(2.25)
where
Here,
formula.
is an imaginary unit and
in accordance with Euler’s
55
Fourier transform on the other hand, transforms a function into a related function of different
variables. This is done by multiplying the original function by a function of both sets of variables
technically called a kernel and then eliminating the first set of variables by integrating the product
with respect to these variables between definite limits. By integrating using the inverse of the initial
kernel, the first variable can be re-obtained. This process is termed inverse Fourier transform. Table
2.1 shows a table of Fourier transform theorems as given by Bracewell (1965).
In signal processing, the Fourier transform often takes a time series or a function of continuous time,
and maps it into a frequency spectrum. That is, it takes a function from the time domain into the
frequency domain; it is a decomposition of a function into sinusoids of different frequencies; in the
case of a Fourier series or discrete Fourier transform, the sinusoids are harmonics of the fundamental
frequency of the function being analyzed.
Discrete Fourier transform (DFT), also called finite Fourier transform is widely employed in signal
processing to analyze frequencies contained in a sampled signal and in operations such as
convolution. Its operation is based on the fast Fourier transform (FFT) algorithm. We use the fast
Fourier transform (FFT) to compute the peak amplitude
FFT,
of the time function
Conversely, given
and
and phase
at every frequency. The
is given by;
, we can synthesize
using the inverse transform
56
Table 2.1.Basic Fourier transform theorems (Bracewell, 1965).
In our equations,
to as transform pairs.
is the FT of
and so,
is the inverse FT of
and they are both referred
is usually complex and so may be separated into real and imaginary parts.
Equation 2.26 then becomes (Telford et al., 1990);
The integrals
are called cosine and sine functions respectively.
Just as the Fourier transform describes the analysis of a time function in its individual frequency
components (Fourier analysis), the inverse Fourier transform synthesizes the time function from the
constituent frequency components. From the appearance of the two transforms it is easily seen that
two successive transformations will yield the original time-function within a constant factor.
Fourier transform operations could be either 1 dimensional or 2 dimensional. Using the onedimensional Fourier transform for a single trace (independent of distance) yields the complex
spectrum as a function of frequency. However, by considering the array of traces only (keeping time
constant), the Fourier transform becomes a function of wave-number only. Obviously, one must
consider both variables, distance and time simultaneously which can easily be achieved by combining
57
the two one-dimensional transforms into one two-dimensional Fourier transform, giving rise to 2-D
Fourier transform. 2-D transform is the basis for both the analysis and implementation of multichannel processes. It is a mathematical process which basically involves transforming a set of data
from time-space domain to the frequency-wave-number domain. The two-dimensional complex
spectrum, therefore, is a function of both frequency and wave-number. 2-D transforms decompose a
wave field into its plane-wave components.
Other types of transform also exist. They are Laplace transform, Radon transform and z-transform.
Transforms and their functions are in domains specified by the variables. They are designed to
preserve information as they switch domains even though minor alterations occur due to
approximations and truncations.
2.6.2. Convolution
Convolution is a process whereby an input time-function is filtered by another time-function to
produce an output time-function. Simply put, it involves multiplying two arrays in a fashion that
results in a third array!
Convolution can be expressed as an operation on sampled data where the unit impulse is denoted by
or as a sample set with a continuous function
. This operation is performed in what is properly
called, the time-domain. Utilizing the Fourier transform, this process can also be described in the
frequency-domain as well. The convolution of two-time functions
dependent function
given by
where denotes the convolution operator.
yields the time
58
Convolution can be defined based on equation 2.29 as the integral of the product of the two functions
after one is reversed and shifted. As such, it is a particular kind of integral transform. The integration
range depends on the domain on which the functions are defined. Using zero-extended or infinite
domains is sometimes called a linear convolution.
However, for sampled functions x(t) and h(t), we simply fold, shift, multiply and add. If the sampled
functions have number of samples
samples
respectively, the convolution y(t) has a number of
given by;
Convolution can be commutative, i.e.
.
If X and Y are two independent random variables with probability distributions x and h respectively,
then the probability distribution of the sum X + Y is given by the convolution
It is usually easier and faster to perform convolution in the frequency domain for functions of long
duration.
Finally, let us consider the convolution theorem. It states that the convolution of two functions in the
time domain is equivalent to multiplying their Fourier transforms in the frequency domain. This is
given by;
The inverse of the convolution theorem is also given as,
59
2.6.3. Correlation
The correlation of two time dependent functions x(t) and h(t) yields the time dependent function as;
The process of correlation is the same as convolution except that the operator is not reversed and
there is no folding around t=0. The output trace is called a correlogram. It aims at comparing traces /
signal and locating the position where similarities exist. In signal processing, the cross-correlation (or
sometimes "cross-covariance") is a measure of similarity of two signals, commonly used to find
features in an unknown signal by comparing it to a known one. It is a function of the relative time
between the signals, is sometimes called the sliding dot product.
When we correlate functions, one of them is usually termed the operator. For sampled functions, we
simply shift, multiply and add. If the sampled functions x(t) and h(t) have number of sample
, the correlation y(t) has a number of samples
given by;
When corresponding values of varying data are multiplied and their product added, what results is
their cross-correlation expressed by Yilmaz (1987) as;
In an Autocorrelation, which is the cross-correlation of a signal with itself, is a measure of how well a
signal matches a time-shifted version of itself, as a function of the amount of time shift and there will
always be a peak at a lag of zero. Autocorrelation is useful for finding repeating patterns in a signal,
60
such as determining the presence of a periodic signal which has been buried under noise, or
identifying the missing fundamental frequency in a signal implied by its harmonic frequencies.
If X and Y are two independent random variables with probability distributions x and h, respectively,
then the probability distribution of the difference − X + Y is given by the cross-correlation of f and g.
In contrast, the convolution, x * h, gives the probability distribution of the sum X + Y.
Applying the convolution theory on the correlation yields;
Correlation is not commutative.
In vibroseis processing, correlation is used to recover a nice pulse from a long out-going sweep (pulse
compression). The uncorrelated seismogram generated from the vibroseis sweep is the convolution of
the sweep with the earth’s response, analogous to equation 2.2. And so, in processing the uncorrelated
seismogram to remove the effect of long sweep, we correlate the vibroseis sweep with the
seismogram to obtain the correlated seismogram.
Cross-correlating recorded data with the oscillatory source itself is very imperative in vibroseis data
processing (Dobrin and Savit, 1988). This involves a comparison of the two signals with
progressively delay times. Related seismic events show greater coherency with source signal than
random noise.
2.6.4. The z-transform
When a continuous function
is sampled, its sampled form
is given as
61
That means, in sampling a continuous function,
by the impulse function every
at a sampling interval
, we simply multiply it
.
In mathematics and signal processing, the z-transform converts a discrete time domain signal, which
is a sequence of real numbers, into a complex frequency domain representation. The z-transform is
used to express the amplitude of successive uniformly spaced time intervals as a polynomial in ; the
power of variable z is the number of unit delays associated with each sample in the series.
Multiplying by
, delays the time series by n time units, while multiplying by
time series by n units. Taking the FT of (2.37), we obtain;
The convolution theory also applies to the z-transform as;
, advances the
62
CHAPTER THREE
LITERATURE REVIEW
3.1. Introduction
The seismic reflection method is a method of exploration geophysics that uses the principles of
seismology to estimate the properties of the earth’s sub-surface from seismic reflected waves. This
method is mainly directed at locating depths of reflecting surfaces and the seismic velocities of subsurface rock layers. It is mainly used for hydrocarbon and mineral exploration due to its ability to
penetrate deep into the earth, up to several hundred meters. Over the years, this method has been
developed to quite a high degree of sophistication.
The method requires a controlled source of energy or signal, usually an explosion or vibration, sent
into the ground at a known time and location. By taking note of the time it takes the signal to be
reflected back from the boundaries between rocks to a receiver on the earth’s surface (usually a
geophone for land and a hydrophone for marine), the depths and features of the sub-surface interfaces
can be estimated after some processing and analyses.
Several field procedures are in use, each being distinguished by the different layout of the detectors
relative to the shot-point.
This report reviews ten (10) reflection operations carried out in different parts of the world, at
different times and for different purposes. The choice of the cases reviewed in this work was carefully
and thoughtfully arrived at. This is to enable a better comparison of data acquisition parameters,
equipment used and the different approaches to data processing and interpretation.
63
The cases reviewed in this report are: Seismic reflection studies in the Skellefte Ore district. Northern
Sweden, Seismic reflection tomography: A case study of a shallow lake in Lake Balaton, COCORP’S
deep seismic reflection profiling in the Williston Basin (comparison of explosive and vibroseis source
energy penetration), Reflection seismic and ground-penetrating radar: Study of previously mined
(lead/zinc) ground, Joplin, Missouri, Seismic processing and velocity assessments in the oil and gas
resource potential of the 1002 area, Arctic National Wild Life Refuge (ANWR), Alaska, Seismic
reflection investigation of C-1 to F-1 sector of the Superconducting Super Collider (SSC) ring west of
Waxahachie, Texas, Analysis of existing seismic reflection data in South Florida for Aquifer Storage
and Recovery (ASR) regional study, 4D acquisition using two boats in the Enfield oil field offshore,
Western Australia,3D interpretation of the Espoir field area, offshore Ivory Coast and Structural and
stratigraphic mapping of Emi field, offshore Niger Delta, Nigeria.
.
3.2. Case 1: Seismic Reflection Studies in the Skellefte Ore District, Northern Sweden
Several world class mines are located in the western part of the Skellefte Ore district, for instance, the
Kristineberg VMS mine. The Skellefte district is a volcanic arc which formed betweena sedimentary
basin (Bothnian Basin) and a continental basin (the Aindsjaur group)(Weihed, 2001). Today,the area
is freely defined (150 by 150km in size) in Northern Vasterbotten and southern Norbotten in the
Northern part of Sweden. The area is rich in mineral occurrence. The surface geology in the
Kristineberg area consists of volcanic rocks overlain by fine-grained turbiditic metasediments.
Data was obtained in this area along two north-south oriented profiles (profiles 1 and 5) between late
August and early September, 2003 (fig. 3.1). Each data was about 25km long.
64
Fig. 3.1 Bedrock geological map of the Kristineberg area, with the locations of the seismic profiles,
the CDP lines and the temporary seismic stations
65
The purpose of the survey was to obtain high resolution images of the upper10km of the crust. The
aim of the survey was to study the very complex crustal architecture of the region in the Skellefte Ore
district in the Kristineberg area at depth and to determine the relationship between the submarine ore
bearing Skellefte group and the overlying volcanic and sedimentary rocks south of the ArcheanProterozoic boundary. Barret and Maclean (2000) has more details on the geology of this area.
Profile 1 was first decided to run parallel to the major tectonic ligaments of the region. The profile
passed on top of the Kristineberg mine and crossed the exposed ore-bearing volcanic units of the
Vargfors group (coarse clastic unit of a sedimentary sequence).
Profile 5 was mainly located along very small roads or sections through rough terrains and runs
parallel to profile 1. This choice was due to lack of infrastructure in the east-west direction and its
ability to allow cross-profile recordings to help correlate observed structures between the profiles.
The separation between both profiles was about 6-8km. Both profiles started in the Revsund granites
in the south, crossed the ore-bearing Vangfors formation and ended in/near the Revsund granites
bordering the Skellefte field to the north (Tryggvason et al., 2003).
Data acquisition:
The 3D reflection survey was conducted within an interval of approximately 3 weeks. Two data
recording systems were used, the SERCEL 348 and SERCEL 408. SERCEL 348, which is an older
version is heavier and more strenuous in rough terrains, was used for profile 1. Also for profile 1, a
group of six 10Hz geophones was used. The more modern SERCEL 408 was used for profile 5
alongside single 28Hz geophones. A summary of the acquisition parameters are shown in table 3.1.
An alternative source of a vibratory system was contemplated but the idea discarded due to very high
mobilization costs. Recording started at the southern ends of the profiles. Data was recorded on
66
temporal seismometers spread out in the area to 20s and sampled every 2ms. GPS instruments used
ensured high precision of shot-hole and geophone locations.
Processing:
PROMAX was the processing package used to process the data which was done at the Department of
Earth Science, Uppsala University. Though recording was done for 20s TWT, processing was done to
4s TWT only (about 12km depth). Similar processing schemes were used for both profiles. Editing
was done in shot domain to remove cultural traffic and ambient noise. Processing sequences applied
to the data are summarized in table 3.2.
Some data on profile 5 were corrected for reverse polarity. First arrivals were picked with an
automatic neutral network algorithm and proper correction effected where necessary after manual
inspection.
3.3. Case 2: Seismic Reflection Tomography: A Case Study of a Shallow Lake in Lake Balaton.
Lake Balaton is the largest lake in Central Europe. It is situated near Budapest in Hungary and is
home to precious natural treasures (fig. 3.2). Lake Balaton is 77km in length and has a width range of
4-14km. Near surface material of the lake consists of unconsolidated, medium to coarse-grained mud
and medium to coarse gravel interspaced with thin discontinuous Paleosols and clay. Test hole of
seismic section in the area indicates a thick Holocene mud layer (3.5 – 4m thick) present just below
the water and a stiff saturated clay stratum just below it (McLamore et. al., 1978).
67
Fig. 3.2. Location map of the broader study area of Lake Balaton.
68
Table 3.1. Data acquisition parameters for profiles 1 and 5
Type of survey
Nominal spread
Nominal fold:
Profile 1
Profile 5
Energy source parameters
Source type
Nominal shot spacing
Charge size
Recording parameters
Number of active recording channels:
Profile 1
Profile 5
Record length
Sampling rate
Maximum offset:
Profile 1
Profile 5
Geophones
Profile 1
Profile 5
Total profile length
2-D Crooked line
End on/shoot through
17
25
Dynamite
100 m
0.5-3 kg
140
200
20 s (about 60 km depth)
2 ms
4000 m
5000 m
Bunch of six 10 Hz
Single 28 Hz
55 km
Data acquisition
The operation was a shallow marine reflection operation. Data was collected using the seismic
profiling system,
receiver with boomer source operating at 150 joules and firing
at 0.25m interval. Table 3.3 shows acquisition parameters.
Acquired data, which was done on a PC-based 50KHz, 16 bit sampling of filtered Seistec data
recorded in SEG-Y format, was done digitally (Soupios, et al., 2005).
Processing
The data was processed on a SUN work station using CWP/ Seismic Unix. In the data processing,
elimination of superimposed surface-related noises on shallow reflections, which is common
especially in shallow marine seismic surveys, consumed most of the processing stages. Such surface-
69
related noises which include direct waves, guided waves, water bottom multiples, reflected refractions
and side scattered noise (Lee, 1999).
Table3.2. A summary of the processing operations on profiles 1 and 5
Read SEG format files
20,000 ms
Reverse negative polarities in
first breaks
Profile5 only
Spike and noise edit
Both profiles
Pick first breaks
Both profiles
Refraction statics
Profile 5: 240 m datum, 1 layer model,
5700 m/s replacement vel.
Profile 1: 300 m datum, 1 layer model,
5500 m/s replacement vel.
Spectral Whitening
Profile 1: Time-variant
Profile 5:
10,20,75,105 Hz
Band-pass filtering
Profile 1: Time & space-variant
Profile 5:
20-40-70-120 Hz, 0-4000 ms
Air blast attenuation
Profile 5 only
Sort to CMP domain
Both profiles
Velocity analysis
Both profiles
NMO correction
40% stretch mute
Residual statics, maximum power autostatics, pass1
Velocity analysis
Residual statics, maximum power autostatics, pass2
Stacking
Both profiles
F-X Deconvolution
Both profiles
Time-variant scaling
Profile 1
Trace equalization
Profile 5
Processing tools applied on the data include band-pass filtering, mute (windowing), f-k filtering,
spectrum analysis, gain deconvolution and migration. Velocity analysis was not possible because of
the single fold nature of the survey, however, a tomographic approach of the model using reflection
travel-times from the pre-defined reflectors was attempted to tackle that issue. In this case, a 5layered, 1D background velocity extracted from a bore-hole in the vicinity of the seismic section was
used. The forward model was solved using finite-difference wave equation modeling.
70
Table 3.3. Acquisition parameters for the Lake Balaton survey
Waveform format
Source
Firing interval
Precision of Positioning
Number of traces
Record length
Sample frequency
Length of profile
Number of channels
Gain
Low cut filter
High cut filter
SEG-Y, 2 bytes integer
Boomer at 150 Joules
0.25 meters
< 1 meter
512
1251 samples
50 kHz
250 m
1
Linear
1 kHz, second order
10 kHz, second order
3.4. Case 3: COCORP’S Deep Seismic Reflection Profiling in the Williston Basin (Comparison
of Explosive and Vibroseis Source Energy Penetration)
In 1990, a joint explosive/vibroseis experiment was scheduled for the fall of that year in the Williston
basin of North Dakota and Montana. However, the dynamite experiment was not performed due to
permit and liability issues. The vibroseis experiment was successful. A key observation from the
vibroseis experiment was the apparent lateral truncation of prominent moho-reflectivities near the
western boundary of the Trans-Hudson Orogen which interpreters thought might be related to phase
changes associated with its formation. Other details of the 1990 survey are summarized in Nelson et.
al., (1993).
In 1992, COCORP (Consortium for Continental Reflection Profiling) assisted by LITHOPROBE,
returned to North Dakota and Montana to collect a new set of explosive/vibroseis data, being guided
by the previous survey of 1990.
The aim of the new survey was to re-image key portions of the 1990 vibroseis profiles with explosive
sources, to collect a north-south profile providing 3D control on Trans-Hudson structure (fig. 3.3) and
71
to confirm the extent of prominent moho reflections as stipulated by the earlier survey.
Fig. 3.3 A map showing the survey area of the Williston basin and Trans-Hudson Orogen of North
Dakota.
72
Another objective of the survey was to test the feasibility of using explosive source in low-fold mode
for deep reflection profiling. The objective of this survey which is of the most interest in this review
is the comparison in the effective imaging of deep structures between the 1990 vibroseis survey and
the 1992 explosive survey.
Data acquisition
Three explosive data were collected in the Williston basin in the 1992 survey; MT12, ND2 and ND3,
each having its own objective. MT12 was designed to test the moho-reflectivity, ND2 was collected
to confirm the along-strike correlation of primary crustal fabrics as suggested by comparison of the
COCORP’s data with subsequent LITHOPROBE surveys in the Saskatchewan to the north (Baird,
1995) and ND3 was acquired to extend imaging further into the bounding superior craton east of the
Trans-Hudson Orogen.
A total of 129 shots were taken for MT12, ND2 and ND3 surveys, of which 97 shots were used to
study shot-to-shot variability. A high fold (30) data was acquired for the vibroseis data while a low
fold (6) data was obtained for the explosive data.
For the 1992 explosive survey, multiple DFS V systems, each equipped with 120 channels recording
system was used. Explosive sources were positioned with a nominal 3km spacing. Explosive shots
varied (from 9.5-90kg) with depths varying from22-38m (140-180ft). This was due to shots being
placed in different media (clay, sand, shale and sand stone). The MT12 survey was shot from west to
east (fig. 3.3) with multiple DFS V systems used in-line to create a 24-30km spread. Nine (9)
L21A(10Hz) geophone was used as detectors.
For the vibroseis survey of 1990, Amoco’s 600+ channel telemetric recording system using radiolinked seismic group recorders deployed in a roll-along configuration was used. The source
comprised eight (8) Mertz vibrators [52,000 lb (23,000N) peak force P-wave] positioned bumper to
bumper to form a 40m in-line linear array. The source array was moved one pad length per sweep
73
with vibration points every 200m. Twelve (12) SM-7 (10Hz) geophones acted as detectors per station.
Production shots for ND2 and ND3 used 45kg of explosive in two shot holes with a separation of
85m.
Processing
It is clearly obvious that the geometries, equipment and processing procedures for both data were
different in every sense. Another major difference was the two years interval separating both
operations. As a result, no quantitative or definitive comparison could be made on both data.
However, given that eleven (11) points from the approximately 35km of vibroseis and explosive data
overlapped on both the western and eastern portions of the transect MT10/12 and ND1-ND3, an
analysis became useful and necessary. Data in the overlapped region included shots and receivers
placed in both surface media (Steer et al., 1996).
Since the acquisition and time of acquisition of both data sets were different, both data were
normalized to minimize amplitude variations caused by the effect of varying ambient noise conditions
and were sealed to reduce synthetic biases which resulted from the use of different recording systems
and geophone patterns. The amplitude normalization was done by taking the absolute amplitude of
samples from the eleven coincident channels centered at 4km (a distance thoughtfully chosen),
summing them up to a single channel and then calculating the RMS amplitudes for selected 100ms
windows from 2-50 TWT.
The correlation process left the vibroseis data bandwidth limited. Although no much difference will
be noticed if the explosive data was band-pass filtered to match the vibroseis bandwidth.
74
3.5. Case 4: Reflection Seismic and Ground-Penetrating Radar Study of Previously Mined
(Lead/Zinc) Ground, Joplin, Missouri.
The Missouri Department of Transport (MoDOT) proposed constructing an inter-state route Alternate
“E” across the Joplin area of Missouri (fig. 3.4). The site had been previously mined for Lead/Zinc
and had been extensively open-pit and so the integrity of the ground became a thing of concern. In the
winters of 1996/1997 and 1997/1998, the MoDOT contracted the departments of Geology and
Geophysics at the University of Missouri-Rolla to conduct a geophysical survey on the entire area to
ascertain the condition of the ground.
The survey had the following objectives;
1.
To map the Mississippian bedrock
2.
To identify and Paleosinkholes and abandoned mine features
3.
To determine the structural geologic trends in the area
4.
To identify major bedrock faults and/or fracture zones
5.
To identify and locate abandoned mine access and ventilation shafts overlain by surficial
milled ore rock.
The applicability of these objectives was primarily to aid MoDOT engineers involved in route
planning, hazard assessment and site mitigation. To achieve this, the University of Missouri-Rolla
conducted two geophysical surveys on the area; a reflection survey to investigate objectives 1- 4
above and a ground-penetrating radar (GPR) survey to attempt achieving the fifth objective.
75
Fig 3.4. Map of the survey area in Joplin, Missouri.
76
Data acquisition
A total of 14,600 lineal meters of shallow reflection seismic data, 15,000 lineal meters of GPR data
and 9 down-hole seismic calibration check-shots were acquired.
An elastic wave generator weight drop was used as source while single channel 40Hz geophones
served as detectors. Seismic reflection data was acquired using a 24-channel Bison Engineering
seismograph with roll-along capabilities. Weight drop source was impacted 6-14 times per shot
record. An interval of 3m was employed for shot, receiver and near-offset spacings. Data processing
was on a Pentium PC using WINSEIS.
Data processing
Processing steps applied to the seismic data include reflection identification from bedrock, muting of
all arrivals prior to bedrock event, first breaks and ground rolls, restoration of shot-gather traces to
CMP gathers, elevation correction, velocity analysis for each line, calculation and application of
surface consistent statics, NMO correction, stacking, residual statics, re-stacking statically corrected
data and band-pass filtering (Shoemaker et al., 2000).
To provide sub-surface lithologic/velocity control and to confirm the interpretation of the reflection
data, MoDOT drilled 23 bore-holes in the area but down-hole calibration check-shot survey data was
acquired at 9 bore-hole locations. Surface to bore-hole tomography was applied with a spacing of 1m.
The check-shot data acquired was used to determine P-wave time-depth and velocity time-depth
relationships for shallow sub-surfaces.
For the GPR profiling, a Geophysical Survey System Inc. sub-surface interface radar unit (SIR-8)
equipped with a 500MHz monostatic antenna was used. A total of 1,500 lineal meters of GPR data
was obtained. The data was acquired as suites of parallel GPR profiles with a separation of 1.5m.
Sampling interval was 30scans/foot and trace length was 100ns.
77
The GPR was processed on a Pentium PC using RADAN. Processing operation applied to the data
include trace normalization, vertical gain application, horizontal and vertical filtering.
3.6. Case 5: Seismic Processing and Velocity Assessments in the Oil and Gas Resource Potential
of the 1002 AREA, Arctic National Wild Life Refuge (ANWR), Alaska.
The Geophysical Service Inc. (GSI), under a contract to a consortium of petroleum exploration
companies in 1984 and 1985, acquired and processed 1,451mi (2,351km) of multi-channel seismic
data on and adjacent to the Arctic National Wildlife Refuge (ANWR). That was part of a petroleum
resource evaluation study which was later performed in 1987 by the Fish and Wildlife Service (FWS),
The Bureau of Land Management (BLM) and the US Geological Survey (USGS). Details of this
evaluation can be seen in (Bird and Magoon, 1987).
At that time in 1987, the USGS reprocessed a part of the seismic data on a VAX11/780 computer
using the DIGICON, Inc., DISCO software package. The reprocessed section was stacked and
intended for use as illustration in the resource assessment publication of 1987.
A renewed interest as to the amount of petroleum that could be buried beneath the ANWR coastal
plain led to a re-evaluation of the resource potential and a re-processing of the 1984/1985 seismic data
with a more modern and more sophisticated workstation in 1996. It was done on a SUN workstation
using PROMAX processing software developed by Landmark, Inc., together with in-house developed
software. This project started in 1996 and the processing was performed on a SUN SPARC Station2
but in August, 1997, the system was upgraded to a SUN ULTRA1because the latter was fast enough
to run some testing of pre-stack depth migration (Lee et al., 1999).
Data acquisition
982.7km (606.6mi) of seismic reflection data using dynamite sources and 247.1km (152.5mi) using
vibroseis source was acquired by GSI from January, 1984 to April 1984. In 1985, another 1121.5km
(692.3mi) seismic reflection data using vibroseis source only was obtained by GSI. A summary of the
78
recording parameters for both dynamite and vibroseis data is shown on table 3.4. Notice the change in
vibroseis sweep for both years and also the change in correlation time.
Data processing
Firstly, the rest of the data was demultiplexed from SEG-B to SEG-Y format and then the geometry
was defined and data stored from shot gather to CMP gathers. At that stage also, the elevation and uphole times or depths were stored in the data base.
A two-window spiking deconvolution with operator lengths of about 180ms was applied to both
dynamite and vibroseis data. That was done to improve temporal resolution. A minimum phase
deconvolution was applied to the dynamite data and a zero-phase deconvolution applied to the
vibroseis data. To correct for elevation, a floating datum was determined. By that floating datum,
elevation statics was computed and applied to individual traces.
F-k filtering was applied on shot domain only even though receiver domain f-k filtering was used in
certain parts of some profiles as an option to improve stack quality. F-k filtering suppresses linear
move-out events and improves S/N ratio significantly.
NMO correction using an automatic stretch with a maximum percentage of stretch of about 20% was
then applied to the data.
3.7. Case 6: Seismic Reflection Investigation of C-1 to F-1 Sector of the Superconducting Super
Collider (SSC) Ring West of Waxahachie, Texas.
Between 19th June and 29th June, 1990, LCT Houston and its sub-contractor, Dobecki Earth Science
(DES) performed a seismic reflection survey along a 16,600ft segment of the SSC, ring west of
Waxahachie (fig. 3.5). The contract was awarded to them by The Earth Technology Corporation
(TETC). That survey followed an earlier feasibility survey carried out some one mile south of the
present survey area in February of the same year (LCT, 1990).
79
Table 3.4. A table showing recording parameters for both dynamite and vibroseis data of the 1002 area
The initial survey was able to define the nature of the Austin Chalk/Eagle Ford shale contact as well
as deeper beds. However, the lack of clarity at the Austin/Eagle Ford (A/E) interface reduced
amplitude at the interpreted base of Eagle Ford and on very deep reflections. This issue necessitated a
more elaborate survey with the major aim of accessing the continuity (faulting) of this key interface
(ie. A/E), which was of major importance to the planning and construction of a ring tunnel.
80
From the earlier feasibility survey, depth to the A/E contact was of the order of 200ft. the thickness of
soils over chalk in the area was much thinner. Typical soil thickness was about 2 inches only, even
though thicker soil and deeper Austin weathering were encountered in some deeper ravines along
several check runs. Surface materials included grass fields, corn fields, treed thickets and water in
cattle stock pond.
The latter survey was decided on to be executed in a similar fashion as the previous. That means same
acquisition, source/receiver type, instrumentation, CDP fold and processing parameters. But due to
the thin nature of the soil coupled with the dry nature of the soil (temperature was typically in excess
of
F), a sledge hammer source was opted for as opposed to the use of ‘Buffalo gun’ source as on
the previous survey. Another reason for this choice was based on the initial noise survey and periodic
comparisons made during the production phase. However, in areas like the northern third of the line
with heavy trees and steep under-growth, a Buffalo gun source was used since it did not require
swinging.
Data acquisition
Data was acquired using a 24-channel seismograph for recording, single 40Hz geophones for
detectors and a sledge hammer (nine stacked sledge hammer blows on a steel plate) as source. Shots
were taken at a distance 60ft from geophone points, geophone spacing was 10ft and consecutive shots
were fired at 20ft interval. The acquired data had a 600% coverage.
81
Fig. 3.5 Location of seismic line segment C1 – F1
82
Processing
The data was processed in the Houston offices of Entropic Geophysics, Inc. (EGI) under the direct
QC supervision of Dr. Thomas, L. Dobecki (DES). The processing sequence applied to the acquired
data is basically the same sequence followed in the processing of the initial data.
Surface consistent residual statics using maximum power of CMP stacks was computed using a
500ms window with multiple gates and a pilot trace containing 31 traces. Though this method worked
well for the coastal plain, however, in complexly deformed areas, a window to guide the static
computation could not be picked, and so, the method didn’t go down well in such areas.
In such cases however, iterative measures were employed, usually 2-3 iterations was required to
obtain the best velocity for stacking. Optimum velocities and statics were used to stack the data.
Several stack methods such as the mean, median and weighted mean were employed in the stacking
process
DMO correction which has an advantage of improving velocity analysis, even velocities for migration
(Deregowski et al., 1981), had been performed on the shot domain of most of the data in the 1987
assessment. DMO velocity analyses were estimated using the same range of CMP gathers. It was
observed in a velocity analysis of sections without and with DMO correction that the DMO velocities
are higher in those sections without DMO correction than those with DMO correction applied to
them.
This can be attributed to the compensation of the dip dependence on velocity performed during the
dip velocity analysis.
Despite that DMO velocities were appropriate for subsequent migration velocity analysis, DMO
stacked sections appeared less inferior to normal stacked sections, as such, majority of the final depth
section came from processed stacked data without DMO correction.
83
A post-stack wavelet deconvolution was applied to all stacked data to correct for differences in source
types and to increase resolution. Prior to the post-stack wavelet deconvolution, a 1,000ms AGC was
applied to the data using a window with operator length 180ms.
A variable norm deconvolution method developed by Gray, (1979) was applied in the wavelet
processing. Different norms, window length and design area were employed in estimating wavelet
between 200 and 1,000ms in the stacked data. About 200 random windows were selected. A single
deconvolution operator was derived from the optimum parameters and applied to the whole line. A
datum correction using simple static shifts, corrected the data to a final flat datum (sea level).
Filtering was done using a band-pass filter with 8-10-68-92 Hz pass-band.
Migration of stacked data was done using Stolt’s wave –number frequency domain migration method.
The method employed migration velocity models derived from the DMO stacking velocities. Prestack depth migration was performed on the data and this produced a clearer image as seen in figure
14, particularly at 10,000ft on the right side of the section.
Depth conversion of the unmigrated data was done using average velocities strictly derived from the
DMO velocities. No well data or VSP data was available in the area to tie to the reflection data.
Velocities were modified where necessary in ascertaining the validity of depth sections. Although the
converted depth section, due to lack of well or VSP data, may lack full certainty, the obtained depth
sections were geologically accurate.
To enhance the visibility of seismic reflections in the data, a 8-12-56-68Hz pass-band was applied by
a median filter using a 3 by 3 grid and their amplitudes were modulated by the trace envelope as
illustrated by Lee et. al.,(1988).
Migration velocity was derived by smoothing DMO velocities using triangular tapers in time and
space. A 300 CMP was used as the half height of a triangular smoother in space and a 500ms was
used as the half height of a triangular smoother with time. Modification of migration velocity was
84
carried out where necessary. Its average velocity was derived using smooth gradient function with the
migration velocity as its input.
3.8. Case 7: Analysis of Existing Seismic Reflection Data in South Florida for Aquifer Storage
and Recovery (ASR) Regional Study.
Several reflection lines are known to exist near and south of Lake Okeechobee in South Florida. Most
of which were collected for oil exploration purposes at depths of about 10,000-14,000ft below ground
surface.
To restore the ecosystem in that region, the Comprehensive Everglades Restoration Plan (CERP)
came up with several projects, one of which was the ASR studies. The ASR studies project was
designed to evaluate the viability and potential effects of constructing about 333 ASR wells in the
area. To do this, a better sub-surface characterization of the area which will optimize effective
location of the ASR became expedient.
Given the numerous seismic lines that traverse the area, a question on the possibility of re-processing
selected portions of available deep reflection data to resolve shallow sub-surface stratigraphy in an
area south of Lake Okeechobee was considered.
In that regard, the US Army Corps of Engineers – Jacksonville district (USACE) teaming up with the
South Florida Water Management District (SFWMD), contracted the URS Group Inc. to review the
assessment in line with CERP’s project.
A seismic data coverage map was obtained from Seismic Exchange Inc. (SEI) for a large area of
Florida. After extensive considerations of the numerous lines on the data, two 5-mile sections, one for
each of the selected lines were chosen. One of the lines utilized a dynamite source with an east-west
orientation and the other utilized a vibroseis south with the same east-west orientation. The selected
85
seismic data were 2D seismic lines acquired in the mid to late 1980s by SEI, both of which targeted
petroleum in very deep zones.
Data acquisition
A total of 399 inputs were recorded for the dynamite line and 179 recorded for the vibroseis line.
CMPs for the dynamite line were 1,157 and 1,055 for that of the vibroseis. The acquisition parameters
for both lines are shown in table 3.5.
Processing
Processing operations applied to the data include demultiplexing of data and converting it to internal
Focus format, trace editing, geometry definition, spherical divergence correction and application of
surface consistent gain. Other processing operations performed on the data are stated in table 3.5.
No elevation data was available for the vibroseis line and so elevations were estimated to be 20ft
MSL at all point. Additional corrections to the data in the CMP domain were done using residual
static programs.
The final migrated stack had a technique known as XFreq applied to it. XFreq is a program designed
by Tricon to boost amplitudes of the higher frequency range thus, enhancing the frequency content of
the data. Its principle is based on dipole filtering. Colton and Nautiyal, (1996) deals more on its
operation.
3.9. Case 8: 4D Acquisition Using Two Boats in the Enfield Oil Field Offshore, Western
Australia.
In 2004, Woodside acquired a baseline seismic data using one boat and dual sources. At the time,
water depth was about 500m and the streamer configuration was 6 by 100m. The data obtained had
good quality with a usable bandwidth of 15-80Hz and the TWT at a target level of 2s. The survey
86
encountered a high degree of feathering which occurs due to current and a change in vessel track
resulting in about 50% infill.
Table 3.5. Acquisition parameters for the ASR studies seismic survey
Vibroseis data parameters:
Number of Recording Channels
Geophone Group Interval
Source Point Interval
Recording Geometry
Data Sample Rate
Record Length
Maximum common Midpoint (CMP) folds
Low Filter
Notch Filter
Alias Filter
Sweep Start
Sweep Rate
Sweep Length
Taper
The dynamite data parameters:
Number of Recording Channels
Geophone Group Interval
Source Point Interval
Recording Geometry
Data Sample Rate
Record Length
Maximum CMP fold
Low Filter
Notch Filter
Alias Filter
480 channels
75 feet
150 feet
Balanced split spread (240-7-240)
4 milliseconds
6,000 milliseconds
120
Out
Out
93 Hz
8 Hz
2 Hz/second
24 seconds
0.5 second
480 channels
60 feet
60 feet
Balanced split spread (240-240)
4 milliseconds
6,000 milliseconds
240
Out
Out
93 Hz
In July 2006, the Enfield field in the north-west shelf of Australia (fig. 3.6) commenced oil
production. Seven months into producing oil in the field, in February 2007, Woodside acquired a push
reverse 4D monitor survey data on the same area, this time using two boats with separate source and
streamer vessels.
87
Table 3.6. A summary of the processing steps applied to the ASR studies data.
Reformat to Focus
Trace Edit
Geometry Definition and QC
Edit for Offsets
Spherical Divergence Correction
Surface Consistent Gain
Surface Consistent Deconvolution
Automatic Gain Control (AGC)
Datum Statics
Velocity Analysis –
Normal Move-out Correction
First Break Suppression
Surface Consistent Residual Statics
Signal Enhancement on Shots
Surface Consistent Residual Statics
Common Mid-Point Gather
Stack
Finite Difference Migration
XFreq
Automatic Gain Control (AGC)
Bulk Shift to Final Datum
The reason behind the choice of two boats was to maximize coverage around obstructions, even
though complete survey using this configuration is rarely achievable and to minimize in-fill as a result
of feathering.
Another key element of the 2007 monitor survey design was the 4D repeatability, maximum offset of
2,300m and the minimal infill due to limited vessel availability (Ridsdill-Smith et al., 2008).
Source repeatability is very high for two boat operation because of the relative smaller source vessels
and easy maneuvering of the source vessel compared to that of a 3D vessel. The push reverse
configuration comprised the source boat following behind the streamer boat, maintaining a safe and
regular distance.
88
Fig.3.6. Location map of the Enfield 4D monitor survey. Inset shows the feather encountered on the
baseline survey in 2004.
89
Data acquisition
The Enfield monitor survey followed the same acquisition parameters as the previously conducted
baseline
survey.
It
followed
the
same
baseline
shot-point
interval
of
18.7
5m (flip-flop), gun separation of 50m and the streamer vessel was steered to match the baseline
receiver positions at a target offset of 1900m.
At the time of the monitor survey, Woodside had only a 4-streamer vessel which couldn’t replicate
the 6-streamer baseline coverage and so, an 8-streamer coverage was used.
3.10. Case 9: 3D Interpretation of the Espoir Field Area, Offshore Ivory Coast.
The Espoir field was discovered in 1980 by a joint venture comprising Philips Petroleum Co., AGIP,
SEDCO Energy and PETROCI. It is located at about 13km (8 mi) offshore Ivory Coast. From
October 1981 to February 1982, GSI carried out a 3D seismic survey which covered 7,700 kilometer
line of data. The area is located on the edge of the continental shelf and extends into deep waters
(fig.3.7).
Before the 3D survey, a previous survey had been carried out on the same area which led to the
drilling of the discovery well A-1X, which was drilled in approximately 515m of water to test a
structural high at the Albian unconformity level. Hydrocarbon was struck.
An appraisal well, A-2X, was also drilled to confirm the presence of hydrocarbon-bearing and
reservoir-quality sands beneath the unconformity surface in the area. A similar exploration in the
adjacent B1 block also revealed interesting features and thus, the need to perform an extensive 3D
survey operation on the area became necessary. The aim of the 3D survey was to obtain a detailed
mapping of the Albian structure.
90
Fig.3.7. Location map of 1981-82 Ivory Coast 3-D seismic survey showing structure of Albian
unconformity (Contours in meters).
91
Data acquisition
The seismic program consisted of 7,700 km line of data containing 525 northeast-southwest-trending
lines acquired in a single survey area located on the edge of the continental shelf and extending into
deep water. Data was recorded using a 2,400m cable and GSI’s 4,000-cu inch air gun source.
3.11. Case 10: Structural and stratigraphic mapping of Emi field, offshore Niger Delta, Nigeria.
Emi field is located in the offshore depobelt (temporally and genetically related families of growth
fault trends or macro-structures) of the Niger Delta and covers an area of 58.24
(fig. 3.8). The
Niger Delta basin is dated as Cenozoic in age and is situated on continental margin of the Gulf of
Guinea in equatorial West Africa. The northern edge of the delta is demarcated by the Cretaceous
Abakaliki anticline, extending further to the southeast as the Afikpo syncline and Calabar flank. To
the northeast of the Niger Delta, is the structurally defined Benue hinge line, while the sediments of
Okitipupa ridge mark the western boundary of the delta (Ugwu and Ezeh., 2012).
The Niger Delta clastic wedge which spans to about 75,000
in southern Nigeria and the gulf of
Guinea, offshore Nigeria, contains reserves of about 34 billion barrels of oil and 93 trillion cubic feet
of gas and results to the 12th largest accumulation of hydrocarbon reserve the world over (Tuttle et al.,
1999).
The deposits of the Niger Delta is divided into three large-scale lithostratigraphic units; the basal
palocene to recent pro-delta facies of the Akata region, the Eocene of recent paralic facies of the
Agbada region and the Oligocene-recent fluvial facies of the Benin formation (Evamy et al., 1978;
Short and Stauble, 1967; Whiteman, 1982)
92
Fig.
3.8. Map of Niger Delta showing the location of the study area (Emi field).
93
In the present work under review, attempts were made to integrate 3D seismic reflection data with
available well logs so as to define the subsurface geometry, stratigraphy and hydrocarbon potential of
Emi field, off-shore Niger Delta. This integration of seismic data with well logs became necessary to
tackle the limitations of well logs in the definition of lateral variation of subsurface parameters. Thus,
the degree of reliability in mapping complex structural and stratigraphic sub-surfaces such as that of
the Niger Delta would be greatly enhanced by combining seismic data with well logs (Barde et al.,
2000; Barde et al., 2002; Adejobi and Olayinka, 1997).
The aims of the study included the characterization of the subsurface geometry and stratigraphy, the
determination of hydrocarbon trapping potential of the field and identification and delineation of
possible hydrocarbon prospects in the field. The composite logs comprised of gamma ray, SP,
resistivities, calliper, sonic and neutron-density logs. Vertical resolution of the physio-chemical
characteristics of the geologic formations of boreholes can be obtained from well logs (Aizebeokhai
and Olayinka, 2011).
Data acquisition
The survey area covered an area of 58.24 sqkm on a scale of 1:250000. Four wells; Emi-1, Emi-3,
Emi-4 and Emi-5 wells were selected for study based on data quality and continuity and depth column
covered by logs. Available data for the study included forty-two 3D seismic sections of which twentyseven of the seismic lines were cross-lines shot parallel to the dip direction and the other fifteen, inlines which were shot parallel to the strike direction. The spacing between the seismic section for both
the in-lines and cross-lines was 400 m. A continuous velocity log of Emi-3 well and composite logs
for six wells and two sidetracks wells which consist of gamma ray, SP, caliper, resistivities and
neutron-density logs were also available for the study.
94
Data processing
Stratigraphic cross-sections were drawn through Emi-3, Emi-1 and Emi-4 wells and Emi-4 and Emi-5
wells to depict the lateral variations of the horizons within the limits of the data provided. Horizons
within the well logs that show hydrocarbon prospect were selected for mapping The selection of
reflection events from the seismic sections was based mainly on amplitude and continuity of such
reflections, especially in areas where there were no well control.
The cross-lines and the in-lines of the migrated seismic sections were tied at their intersection points
and seismic reflection events that show reasonable amplitude and continuity were selected while areas
with closures or misties (attributed to soft shales) were rechecked and analyzed. Major faults were
identified based mainly on break in reflection events or abrupt termination of reflection events and
marked on the cross-lines. The fault traces were posted to the structure contour maps produced.
The thicknesses of these horizon were converted to two-way travel times using the time-depth relation
curve. The thicknesses of these horizon were converted to two-way travel times using the time-depth
relation curve. The faults’ throws and directions were obtained from the seismic sections and
correlated with well logs where there were well controls.
At the end, four horizons, H1, H2, H3 and H4 were mapped and structure contour maps produced for
each of the horizons. To obtain depth information for horizon H4 (the deepest of the horizons
selected), the time-depth relation curve was extrapolated to a two-way time of about 2700 ms.
After digitization of the selected horizons using appropriate scales, a structured contour map was
constructed for each.
The gamma ray, SP and resistivities logs were used in identifying the lithologies (within the Agbada
Formation) penetrated by the wells. The resistivities logs and neutron-density logs were used in
distinguishing between saline-bearing water formations and hydrocarbon pay zones. Consistency was
ensured in all the data sets.
95
CHAPTER FOUR
RESULT DISCUSSION
4.1. Results of Case 1
The data acquired as described in the previous chapter, each underwent different acquisition and
processing operations and as such, one would expect different results. This chapter intends to
compare the results obtained in the cases discussed in chapter three.
For profile 1, a constant velocity analysis shows that the best stacking velocity for this profile is in the
order of 11,000m/s. Bearing in mind that the stacking velocities for dipping events depend on the dip
of the reflector, imaging clearly, all dip directions in one location tends to be a difficult task. To
correct for this, a DMO correction was applied to the data set. After which, reflections with
conflicting dips were successfully imaged with stacking velocities in the order of 6,500m/s.
A brute stacked section of the data (fig. 4.1a) before and (fig. 4.1b) after the application of refraction
and residual statics correction as well as time and space-variants band-pass filters can be seen in
figure 4.1. The band-pass filter enabled clearer coherent reflections as seen in the comparison of both
displayed sections. Strong reflectivity can be seen after 2s in (b), dipping northwards between CDPs
1,000 and 1,800 and series of sub-horizontal reflections are imaged between CDPs 90 and 800,
starting at 2.5s. Several clear reflections can be observed within the region but a very interesting
feature is an apparent south dipping that appears to approach the survey very close to the Kristineberg
mine.
96
(a)
(b)
Fig. 4.1.A brute stack showing (a) before and (b) after applying refraction and residual statics, and a
time and space-variant band-pass filter. The reflectivity is much more coherent.
For profile 5, figure 4.2 shows (a) an unmigrated and (b) migrated line-drawing of a 4s TWT of
profile 5 after processing. Figure 4.2a shows several reflections most of which can be traced to the
surface of which, most information deduced from it can be misleading. Figure 4.2b shows a better
image of the dip of the reflectivity. For instance, figure 4.2a shows B1 and B2 dipping opposite to B3
97
and B4 and another strong event G1 to G2, visible from 1s to about 3.5s, dipping to the north, and
both (i.e. B and G) crossing each other. Figure 4.2b clearly reveals that both reflective packages do
not cross.
Due to the complex geologic setting of the Kristineberg area and the need to analyze the 3D
orientations of reflections observed on stacked section on profile 5, the strike and dip of some of the
reflectors were modeled. Velocity was assumed to be fairly homogenous and the reflector was
allowed to be a plane with any orientation in space instead of occurring in the line of the profile.
4.2. Results for Case 2
Figure 4.3 shows the raw zero offset section of the data acquired in Balaton Lake. Three distinct
reflections can be observed. Also observable is a trapped gas bubble visible at about trace 430.
Figures 4.4a-d shows the same data after (a) band-pass filtering, (b) gain recovery function, (c)
deconvolution (optimum Wiener deconvolution) and (d) phase-shift migration of zero-offset data
based on the work of Gazdag and Sguazzero (1984) have been applied to it. A significant
improvement in data quality can be observed. The reflection tomography approach provided better
control of the configuration of the reflecting interfaces so that all that was done was to invert for
reflector geometries. Figure 4.5 is a tomogram showing the position of the reflectors before (dashed
lines) and after (thick lines) inverting reflection travel-times.
The final tomographic solution suggests a shallower reinstatement of the initial interfaces, especially
the first and second interfaces, as described by a flat reflector at depths 3-9m and 8m respectively.
The third reflector is a dipping interface with depth starting from 14m to 9m at the end of the
tomogram.
98
Fig.4.2. Line drawing of the stacked section along Profile 5 (a) before and (b) after migration and
depth conversions. B1 to K1 denote reflective segments.
99
Fig. 4.3. Raw zero offset section (512 traces with 25cm trace spacing) containing three reflectors
acquired in Balaton Lake.
4.3. Results for Case 3
Comparing the stacked seismic sections for the COCORP Williston basin survey, that is, the high fold
(30-fold) vibroseis data with the nominally low fold (6-fold) sections from the explosive data (fig.4.6)
reveals the superiority of the vibroseis data in imaging the shallow sections (0-5s) over the explosive
data. The vibroseis data, due to its high redundancy and more uniform source-to-group coupling,
facilitated static corrections, enhanced stacking and provided adequate energy penetration to at least
moho travel times in the thick cratonic crust.
However, explosive data was observed to provide a more superior imaging of the lower crust on a
shot-to-shot basis with amplitudes above ambient noise levels to much greater travel-times. As
such, the explosive source has a higher penetration power through the sediments of the lower crust
than the vibroseis. Nevertheless, size of charge proved to be irrelevant in the imaging of the lower
and upper crust as a single , well-placed, moderately sized (30kg) charge was as good in penetrating
to mantle travel times as larger (90kg) charges.
100
(a)
(c)
(b)
(d)
Fig.4.4. The same data in figure 4.3 after (a) band-pass filtering, (b) gain recovery function, (c)
deconvolution (optimum Wiener deconvolution) and (d) phase-shift migration of zero-offset data.
This can be seen in the relative lack of amplitude variation with increasing or varying charge size for
travel-times beyond 35s in the region.
101
A technical comparison between both surveys, confirmed that there is no distinct moho underlying
the Trans-Hudson Orogen in the region. The relative lack of amplitude variation with increasing or
varying charge size for travel times beyond 35s in the region suggests that moderately sized sources
were as effective as larger sized ones.
Fig.4.5. Reconstructed reflectors after inverting reflection times and appropriate regularization
factors have been applied. Dashed lines represent initial reflectors and thick continuous lines
represent the final inverted geometry.
Further deductions from the comparison of both sources showed that low-fold source-to-ground
coupling, large offsets spanned by CDP gathers (30km versus 15km for vibroseis), ray-path
distortions across large spread and large distances between explosive shots and source-to-source
variability could be problematic to explosive data acquisition.
It was also confirmed that there is the presence of a dipping reflection fabric in the upper mantle
along the western flank of the Trans-Hudson Orogen buried beneath the Williston basin.
102
Fig. 4.6 A stacked section of data obtained from lines (a) MT10(vibroseis) and (b) MT12 (explosive)
imaging the same unmigrated sub-surface features through 30s TWT. (c) is still MT12 imaging the
upper mantle features.
4.4. Results of Case 4
Correlated reflections from the Mississippian bedrock with profiles from the seismic data
corresponded consistently with each other. Bedrock depth variation was observed to be 3-33m which
was the reason for the limited bore-holes in the area. Pennsylvanian karst-collapse features
(paleosinkholes) were seen as structural lows and appeared elongated, bounded by fault scarps. They
were developed along solution widened fractures and were denoted as high risk zones due to mining
103
activities performed on the area. Low risk areas are those areas that were structurally elevated and
were less likely to be mining sites.
The GPS profiles crossed undisturbed clay-rich residual soil overlain by a thin veneer of chat. The
chat/soil contact was easily identified and correlated across the GPR profile. Those profiles imaged
in-fill mine access shafts. The result suggested that materials had been displaced possibly by a
bulldozer to fill local drive way surfaces, as much of the chat had been illegally gathered away. The
chat/soil interface is regular in character and continuous except where the GPR profile crosses the
mine access or ventilation shaft.
4.5. Results of Case 5
The USGS at the time, lacked software packages to interpolate automatically and smoothen all the
average velocity analyses generated in the ANWR seismic re-processing into a 3D model. If it
weren’t so, an improved second order anomaly in the velocity model would have been obtained.
However, only one 3D velocity model for the top of the pre-Mississippian meta-sedimentary rock
horizon could be obtained in the survey.
The re-processing of the seismic data derived in 1984/85 for the 1998 assessment in 1996/97 was able
to improve significantly the temporal resolution of the data over the area. Consistent wavelets were
generated through the wavelet processing which merged satisfactorily to improve data quality. By
merging appropriate lines also, the edge effect of the migration and velocity field was reduced. This
also improved image quality of the data.
4.6. Results for Case 6
The results obtained showed the same basic reflection as in the initial survey although the image
clarity differed as a result of the different sources used, different ground conditions and differences in
depth to A/E. The very dry ground in the latter survey resulted in the less effective power of both the
104
hammer and Buffalo gun sources at higher frequencies when compared to the former, thus, reducing
energy seen at depth and consequently, their reflection amplitude.
Data from both surveys were observed to tie well in areas where the depth to A/E is similar and show
poor events in areas where the depth to A/E is at 50% of its depth at the other side.
The very shallow parts of the data was degraded by noise generated at the shot-point and so produced
fewer usable traces. Though degraded in quality, the obtained sections reproduced the desired crosssectional information which includes the A/E interface, intra-Eagle Ford reflection, the base of the
Eagle Ford and the strong deeper reflections from below the Eagle Ford.
The processing operation was able to remove noise, especially surface waves, increase data resolution
and adjust for topographic and irregular weathering-related time shifts. It was also able to combine
separate results which increased signal levels.
The interpretation was based on the continuity of all cross-sectional information and how they tie to
sonic logs and shallow bore-hole data along the line. By following the continuity of a specific
reflection, the dip changes and/or fault offset of that bed can be determined.
The data was structurally interpreted to have significant disruptions in a few zones along the line.
Three zones of possible faulting were interpreted. These zones were generally grabens or normal
faults with displacements of about 20ft. Those fault features were fairly distinct and separate. There
was no clear tie between faults at depth with shallower faults which cut the Austin.
4.7. Results for Case 7
An obvious improvement was observed on the stacked section, the migrated section and the stacks on
which the enhanced frequency technique was applied, with the latest showing better reflection
continuity and resolution. The migration process significantly reduced scattering of reflections.
105
Higher frequencies were observed in the explosive data even though the vibroseis sweep was limited
to 56Hz. However, the vibroseis data contains less significant gaps which were more obvious in the
explosive data.
Both the explosive and vibroseis lines suggested good resolution of reflections. While the vibroseis
data displayed good resolution from approximately 200ms (depth of 500-600ft), the explosive data
showed good resolution at about 100ms (minimum of 400ft).
On several portions of the explosive line, observed anomalous features suggested potential buried
channel-like sequences, faulting and/or fracturing as well as several potential collapse structures.
While on the vibroseis line, anomalous features were detected as well as potential collapse structures.
4.8. Results for Case 8
Feathering or cross-drift of streamer and sources which arose mainly due to currents and changes in
vessel track was the major problem encountered in the one-boat baseline survey which necessitated
the use of two boats to curb the effect. The latter survey significantly suppressed that effect.
A good positional repeatability in the monitor survey with a mean
error of 27m at the
target offset of 1900m was achieved with the two-boat operation. Desired far offsets were accurately
targeted by decoupling the sources and receivers unlike in the one-boat case where good repeatability
is assured at near offsets if the vessel is steered to the source track and far offsets are reasonably
affected by feather mismatch. However, the receiver motion reduced the repeatability of the two-boat
monitor survey because it moved against the water current during the recording of each shot.
4.9. Results for Case 9
The 3D data obtained was processed by the GSI and migrated in 3D. Bulk of the mapping was based
on a combined interpretation of vertical sections and time slices. The latter was most useful in areas
with poor lateral resolution. Figure 4.7 is a time slice well beneath the Albian unconformity. The red
bands mark the traces of major faults whose individual seismic characters can be identified. A close
106
look at the top part of the figure however, indicates a zone of poor data quality resulting to fault traces
not being adequately mapped. In such cases, closely spaced (60m) vertical sections were used to
define fault traces. They were also used to correlate weak or complex reflections and to map smaller
depositional units. An improved velocity derived from the closely spaced velocity analysis was used
to prepare depth maps at the reservoir level.
Data obtained from the 3D survey when compared to the previous 2D survey done on the same area,
aided interpretation and mapping by providing sharper definition of structural features, showing more
reliable correlation of horizons and fault traces between closely spaced tracks. It also aided in the
preparation of detailed time contour maps from time-slice sections and an improved velocity depth
model for depth conversion. In defining structural features, a comparison can be made of a migrated
section of the 2D and 3D data. Figure 4.8 shows the two versions of line 358 which crosses a feature
near A-4X. Although at about time 2.7s, an evidence of anomaly can be seen at the centre of the
figure but the feature itself is not clear. On the same data after 3D migration (right), a more improved
detail is observed. It shows a steeply-dipping intra-Albian reflection cutting through a flat spot which
is close to a fluid contact defined in well A-4X.
4.10. Results of case 10
The stratigraphic cross-section through Emi-4 and Emi-5 wells tended to be more horizontal and
parallel to one another than those of the cross-section through Emi-4, Emi-1 and Emi-3. This is
because Emi-4 and Emi-5 are more or less in the same strike direction compared to Emi-4, Emi-1 and
Emi-3.
107
Fig.4.7. In-line 525, crossing A-1X well location, Espoir field, and showing clear definition of rotated
fault blocks beneath Albian unconformity.
108
The identification of lithofacies on the seismic sections was based primarily on the amplitude of
seismicreflection events. The high amplitude and continuous reflection were found to correspond to
sand units, whereas the low amplitude reflections were found to correspond to shale facies. Poor
continuity and relatively low reflections were attributed to shaly sand or sandy shale units.
Primarily, twenty seven (27) sand and shale units of varied thicknesses were observed. Most of the
units are laterally continuous, though few cases of pitch-out/wedge-out are evident.
Variations were observed in the thicknesses of horizons especially in Emi-1 where the reservoir units
were thickening and the shale facies, thinning. The variations observed in lithofacies thicknesses
could be attributed to variation in sediment supply, rate of sea level rise and fall, paleogeomorlogy,
synsedimentary tectonism or error in data processing.
Cross-sections through its structure maps also suggest a system of growth faults, roll over anticlines
and folding. Most of the major faults picked on the seismic records were counter regional growth
faults which are characteristics of the self edge, offshore Niger Delta (Ojo, 1996). These counter
regional faults trended west-east and dipped south-east.
Two regional growth faults were picked which had throws ranging from 46 to 79 m (150 to 260 ft) for
the major faults, while that of the minor faults ranged from 24 to 36 m (80 to 120 ft). This throw is
appreciable and could have produce migration pathway for hydrocarbon.
The structure contour maps of the horizons mapped, H1, H2, H3 and H4. The horizons mapped are all
within the Agbada formation, where most of the hydrocarbon is believed to be trapped in the Niger
Delta. The contour maps showed a system of rollover anticlines associated with growth faults. The
faults appeared crescent-shaped with the concave side being towards the down-thrown block.
109
Fig.4.8. Comparison of 2-D migration and 3-Dmigration sections across structure drilled by well A4X showing improved definition of erosional high on Albian unconformity and fluid contact (flat
spot).
Faulted and folded anticlinal closures were common structures observed in all horizons. These
closures serve as good traps for hydrocarbon and are therefore possible hydrocarbon prospects. Apart
from delineating the structural traps, other stratigraphic features such as pinch-outs, unconformities,
sand lenses and channels were also suspected.
110
CHAPTER FIVE
DISCUSSIONS AND CONCLUSION
5.1. Discussion and conclusions
The successful interpretation of a seismic data, be it 2-D/3-D or land/marine, depends on proper
survey design, acquisition and processing, guarding against pit-falls and optimal utilization of
available technology. The cases reviewed in this work each had different objectives and as a result,
underwent different processing approaches.
This review has been able to demonstrate the usefulness and limitations of certain approaches to
seismic operations. It has also been able to establish the fact that seismic reflection method remains a
very vital and indispensable tool in the analysis and structural mapping of the sub-surface.
Today, more and more advancements have been recorded in the exploration industry, such
advancements include 4-D seismic survey (time-lapse), multi-streamer and multi vessel survey for
marine operations and enhanced oil recovery (EOR) which are current trends in seismic survey and
are gaining popularity in the area of reservoir evaluation, prediction of future productions and
hydrocarbon recovery. There is today an increasing interest in the other forms of energy such as Swaves that arrive at the surface spread other than P-waves which are only part of the wave field.
These other forms of energy may be mode converted from a down-going impulsive source or may
arise from a shear wave source capable of outputting both compressional and shear waves chances
are that more developments will be achieved having better of these waves.
Technological advancements which have also richly enhanced interactive software allow users to
iterate parameters and view effects on screen almost immediately in animated form. Such software
111
have been integrated with measurements, analyses, interpretation and reservoir properties to help the
interpreters work better.
The reviewed cases have been able to demonstrate that it is possible to determine the strike and dip of
reflectors using the crooked line geometry.
It was also able to prove that the tomographic approach is capable of providing depth models even in
the absence of bore-hole control and velocity analysis. On the other hand, the tomographic approach
was not able to detect and define the biogenic gas bubbles present in the raw seismic data even though
it was able to locate the anomalous points. This was due to the relatively large wavelength used.
Another reviewed case of the Niger Delta of Nigeria was able to demonstrate that the integration of
seismic data and well logs is a useful and valid tool in structural and stratigraphic mapping and in
better definition of the subsurface geometry and hydrocarbon trapping potential.
The usefulness and limitations of vibroseis and explosive sources for deep crustal and upper mantle
exploration was also revealed in the profiling of the Williston basin. The vibroseis data produced a
more uniform source to group coupling and also a better image of the upper crust while the explosive
data, due to its high penetration ability gave a better image of the lower crust.
It also established the fact that reflection seismic survey can be used to map bedrocks and bedrock
related features such as faults and paleosinkholes and also to map boundaries of previous mine sites.
A knowledge of faults and shear zones, sinkholes etc, helped avert potential highway and/or
construction hazards and aided route planning.
The re-processing of the seismic data derived in 1984/85 for the 1998 assessment in 1996/97, was
able to improve significantly the temporal resolution of the data over the area. This goes to buttress
the improvements recorded in the industry as a result of technological advancements. Today, we have
112
even more improved software some of which include Petrel,
,
,
,
to mention but a few.
We also re-established the fact that data from a small source (sledge hammer) could also be used to
image the upper crust with as much clarity as that of a large source if there is proper coupling of the
source to the ground and adequate processing sequences applied.
The use of modern software and hardware in data processing as elaborated in the seismic processing
and velocity assessments in the oil and gas resource potential of the 1002 area emphasizes on better
processing operations using more modern equipment. Much emphasis was placed in the ASR regional
study case on XFreq, a modern software (at that time), which boosted amplitudes of the higher
frequency range and thus produced a more enhanced frequency content of the data.
It also shows that explosive data have higher frequency content than vibratory data and as such would
image deeper layers better.
For marine operations, multi-streamer surveys curb such limitations such as feathering which was
inherent in the reviewed one boat survey, provides a good positional repeatability and allows better
coverage of the survey area.
The superiority of 3D over 2D survey was also demonstrated in this review. The 3D survey data
brought about a better definition of pre-unconformity reflections which resulted in better mapping of
the intra-Albian horizons and better correlations across major faults. Such improved mappings aided
in the identification of additional wells.
5.2. RECOMMENDATION
In line with the conclusions reached from the reviewed seismic reflection operations, we hereby
recommend that adequate caution should be taken against pitfalls and errors especially during the
113
acquisition and processing of seismic data, since any error in either regard will be quite misleading
during interpretation.
We also recommend that acquired seismic data should be integrated with available well logs to
provide a more reliable and better definition of the sub-surface.
We were able to see that dynamite sources provided data which gave a better definition of the deeper
crust than vibroseis and other sources and as such, we recommend a well uniform source-to-ground
coupled dynamite source for deep crustal surveys and a well impacted uniform vibroseis or any other
(boomer, hammer etc) source for shallow surveys.
To curb the effect of feathering and poor positional repeatability in marine surveys, multi-streamer
vessel survey is recommended.
In seismic data acquisition and processing, we recommend the use of more modern equipment
(hardware and software).
Lastly, for a more intensive and more detailed stratigraphic and structural definition of the earth crust
be it land or marine, we recommend a 3D survey over its 2D counterpart.
114
REFERENCES
Adejobi, A. R. and Olayinka, A. I., (1997). Stratigraphy and hydrocarbon potential of the Opuama
channel complex area, western Niger delta. Nigerian Association of petroleum Explorers
(NAPE) Bull., 12: 1-10
Aizebeokhai, A. P. and Olayinka, I., (2011), Structural and stratigraphic mapping of Emi field,
offshore Niger Delta, Nigeria. Journal of Geology and Mining Research Vol. 3(2), pp. 25-38,
February 2011.
Baird, D. J., (1995). Trans Hudson Oregen structure beneath the Williston Basin results from
COCORP seismic reflection profiling: Ph.D thesis, Cornell University
Barde, J. P., Chambertain, P., Gralla, P., Harwijanto, J., Marsky, J., Schroeter, T., (2000). Explaining
a complex hydrocarbon system in the PermoTriassic of the precaspian basin by integration of
independent models. Abstracts, 62nd European Association of Geoscientists and Engineers
Conference and Technical Exhibition, 2: (P.021), P. 4
Barde, J.P., Gralla, P., Harwijanto, J., Marsky, J., (2002). Exploration at the eastern edge of the
prescapian basin impact of data integration on Upper Permian and Trasicprospectivity.
American Association Petroleum of Geologists Bulletin, 86: 399-415.
Barrett, T.J., Maclean, W.H., (2000); Chemostratigraphy, Petrography and Alteration at the
Kristineberg VMS Deposit, Skellefte District, Sweden. Unpublished report for Boliden AB
Exploration.
Bird, K. J. and Magoon, L. B., (1987); Petroleum geology of the northern part of the Arctic National
Wildlife Refuge, northeastern Alaska: US Geological summit bulletin 1778, P. 329.
Bracewell, R. N. (1965), The Fourier transform and its application:
McGraw-Hill Book Co. New
York.
Brown, A. R., Grillot, L. R.,Anderton, P. W.,Haselton, T. M. andDermargne, J. F., (2004); 3D
Interpretation of the Espoir Field Area, Offshore Ivory Coast, Interpretation of ThreeDimensional Seismic Data, Sixth Edition; AAPG Memoir 42 SEG Investigations in
115
Geophysics, No. 9; The American Association of Petroleum Geologists and the Society of
Exploration Geophysicists, Tulsa, Oklahoma, U.S.A.
Colton, P. B. and Nautiyal, A., (1996); Cascaded Dipole Filters: Extending the Limits of Seismic
Resolution, CSEG Recorder, P. 8-19.
Constain, J. K. and Coruh, C., (2004), Basic theory of exploration seismology: Elsevier press, Oxford,
U.K.
Deregowski, S. and Rocca, F., (1981); Geometrical optics and wave theory for constant–offset
sections in layered media;, Geophysical prospecting, v. 29, P. 374.
Dobrin, M. B. and Savit, C. H., (1988), Introduction to geophysical prospecting: McGraw-Hill, New
York.
Evamy B. D., Herebourne J, Kameling P, Knap W. A., Molley, F. A., Rowlands P. H., (1978).
Hydrocarbon habitat of Tertiary Niger Delta. American Association of Petroleum Geologists
Bulletin, 62: 1-39.
Gazdag, J., and Sguazzero, P., (1984); Migration of Seismic Data by Phase Shift Plus Interpolation,
Geophysics, 49 (2), 124-131.
Grant, F. S. and West, G. F., (1965), Interpretation theory in applied geophysics: McGraw-Hill, New
York.
Gray, W. C., (1979); Variable norm deconvolution: Palo Alto, CA, Stanford University Ph.D. thesis,
P. 101.
Hatton, L., Worthington, M. H. and Makin, J., (1986), Seismic data processing (Theory and
Practice): Blackwell Scientific Publications, London.
Keary, P. and Brooks, M., (1984), An Introduction to geophysical exploration: Blackwell Scientific
Publication, England.
LCT Houston, Inc. and Dobecki Earth Sciences Houston, Texas., (1990); Seismic Reflection
Investigation of C-1 To F-1 Sector of the Superconducting Super Collider (SSC) Ring West of
Waxahachie, Texas.SSC-SR- 1137, Project No. 87-888-0022,Report No. SSCGR-78,
September 1990.
116
Lee, M. W., Agena, W. F., Grow, A. J. and Miller, J. J.,(1999); The oil and Gas potential of the 1002
area, Arctic National Wildlife Refuge, Alaska, by ANWR assessment team, U. S. Geological
Survey Open File Report 98-34.
Lee, M. W., Agena, W. F. and Hutchinson, D. R., (1988); Processing of the GLIMPCE MultiChannel Seismic Data: US Geological Survey Open File Report 88-225, P. 46.
Lee, W. M., (1999), Processing of Single Channel Air Ad Water Gun Data for Imaging an Impact
Structure at the Chesapeake Bay, U.S. Geological Survey Bulletin 2169, U.S. Department of
the Interior.
Lowrie, W., (2007), Fundamentals of geophysics: Cambridge University press, New York.
McLamore, V.R., Anderson, D.G., and Espana, C., (1978), Crosshole testing using explosive and
mechanical energy source, in Dynamic Geotechnical Testing. ASTM STP 654: ASTM, 3050.
Merki, P. J., (1972). Structural geology of the Cenozoic Niger Delta: 1st Conference on African
Geology Proceedings, Ibadan University Press, pp. 635-646.
Murat, R. C., (1970). Stratigraphy and Paleogeography of the Cretaceous and Lower Tertiary in
Southern Nigerian. 1st Conference on African Geology Proceedings, Ibadan University Press,
pp. 251-266.
Nelson, K. D., Baird, D. J., Walter, J. J., Hauck, M., Brown, L. L., Oliver, J. E., Ahern, J. L., Hajnal,
Z. and Sloss, L. L., (1993); Trans Hudson Orogen and Williston Basin in Montana and North
Dakota: New COCORP deep-profiling results: Geol., 21, P. 447-450.
Ojo, A. O., (1996). Pre-drill prospect evaluation in deep water Nigeria. Nigerian Association of
Petroleum Explorers (NAPE). Bull., 11: 11-22.Ojo, A. O., (1996).
Ridsdill-Smith, T., Flynn, D. and Darling, S., (2008); Benefits of two-boat 4D acquisition, an
Australian case study; The Leading Edge, July 2008 edition, P. 940-46
Robinson, S. R. and Coruh, C., (1988), Basic exploration geophysics: John Wiley and Sons.
Sengbush, R. L., (1983), Seismic exploration methods: IHRDC, Boston.
117
Shoemaker, M. L., Anderson, N. L., Hatheway, A. W., Newton, E. T., Cardimona, S., Shaw, A. E.,
Baker, A. J., Webb, D. and Conley, J., (2000);Reflection Seismic and Ground-Penetrating
Radar Study of Previously Mined (Lead/Zinc) Ground, Joplin, Missouri Department of
Transportation Special Publication (2000).
Short K. C., Stauble A. J., (1967). Outline of Geology of the Niger delta. American Association of
Petroleum Geologists Bulletin, 51: 761-779.
Soupios, P. M. and Toth, T., (2005); Seismic Reflection Tomography: A Case Study of a Shallow Lake
in Lake Balaton. Journal of Balkan Geophysical Society, vol. 8, no.1, February, 2005, p. 2027
Soupios, P.M., Papazachos, Tsokas, G.N., Vafidis, A., (2004), Nonlinear Simultaneous
Determination of Reflector Geometry and 3-D Velocity Structure, in review in Geophysical
Prospecting.
Steer, D. N., Brown, L. D., Knapp, J. H. and Baird, D. J., (1996); Comparison of Explosive and
Vibroseis Source Energy Penetration During COCORP’S Deep Seismic Reflection Profiling
in the Williston Basin. GEOPHYSICS, Vol. 61, NO. 1 (January-February 1996); P. 211-221
Telford, M. W., Geldart, L. P. and Sheriff, R. E., (1990), Applied Geophysics (Second Edition):
Cambridge University Press, New York.
Tryggvason, A., Malehmir, A., Rodriguez, J., Juhlin, C., Weihed, P. and Palm, H. (2003); Seismic
reflection studies in the Skellefte Ore District, Northern Sweden – towards a 3-D geological
model, Annual Report 1, 2003-2004
Tuttle, M. L. W., Charpentier, R. R. and Brownfield, M. E.,(1999), The Niger delta petroleum system:
Niger delta province, Nigeria, Cameroon, and Equatorial Guinea, Africa: USGS Open-file
report 99-50-H.
Ugwu, G. Z. and Ezeh, C. C., (2012), Structural and stratigraphic mapping of Emi field, offshore
Niger Delta, Nigeria. International Research Journal of Geology and Mining (IRJGM) (22766618) Vol. 2(7) pp. 199-204, September 2012.
Weihed, P., Bergman, J., Bergström, U., (2001). Metallogeny and tectonic evolution of the Early
Proterozoic Skellefte district, northern Sweden. Precambrian Research 58, 143-167.
118
Whiteman, A. J., (1982). Nigeria: Its Petroleum Geology, Resources and Potentials. Graham and
Trotham, London, 2, 306-361.
Yilmaz, O., (1987), Seismic data processing: Society of exploration geophysics, Tulso Okla.