Download Title Seismic modeling with Discontinuous Galerkin Finite

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Choice modelling wikipedia , lookup

Transcript
Title
Seismic modeling with Discontinuous Galerkin Finite-Element method: application to large scale 3D
elastic media
Author
V. Etienne, J. Virieux, S. Operto
Abstract
We present a Discontinuous Galerkin (DG) finite-element formulation with convolutional perfectly matched
layer (CPML) absorbing condition suitable to large scale seismic modelling in 3D elastic media. The
issues of performance and load balancing of the proposed scheme are discussed in details. A special
attention is given to the last point, a critical issue in applications based on a parallel architecture with
MPI communications like most of geophysical applications. When dealing with large scale modelling, it
is of first importance to seek for an efficient and inexpensive absorbing condition. Thanks to the mix of
interpolation orders which is a key feature of the DG method, we get a good compromise by combining
the DG P1 scheme in the medium to get a acceptable accuracy and the DG P0 scheme in the absorbing
layers to reduce the whole numerical cost. A good load balancing is observed between the processes
without paying any effort on the subdomain decomposition.
71st EAGE Conference & Exhibition — Amsterdam, The Netherlands, 8 - 11 June 2009
Introduction
The simulation of wave propagation in complex medium has been efficiently tackled with finite-difference
(FD) methods and applied with success to numerous physical problems since the past decades. Nevertheless, FD methods suffer from some critical issues inherent to the underlying Cartesian grid such as
parasite diffractions in case of boundaries with complex topography. Therefore and thanks to the ever
increasing computation power, others methods have focused a lot of interests. Concerning the 3D elastic
seismic wave equation, we can mention the Spectral Element method (SEM) popularized by Komatitsch
and Tromp (1999) and more recently the Discontinuous Galerkin (DG) method (Dumbser and Käser,
2006) which has been proved to give accurate results on tetrahedral meshes.
Here, we present a DG formulation with the convolutional perfectly matched layer (CPML) absorbing
condition (Komatitsch and Martin, 2007) suitable to large scale 3D seismic simulations. In this context,
the DG method provides major benefits. Thanks to tetrahedral meshes, one can fit almost perfectly complex topographies or geological discontinuities and the discretisation can be adapted locally to media
properties. Moreover, the DG method is completely local and therefore suitable for parallelization. We
should also mention the possiblity to mix different interpolation orders. The main drawback of the DG is
its relatively high numerical cost due to the fact that elements do not share their nodal values contrary to
classical continuous finite-element methods. In the following, we take benefit of the mix of interpolation
orders to achieve an absorbing condition with a reduced numerical cost by using low order in the absorbing layer. In the next sections, we introduce the DG method with the CPML formulation, we discuss
the computing aspects regarding the proposed scheme and we illustrate the efficiency of the proposed
method with a target of the SEG/EAGE overthrust model.
DG formulation with CPML
We apply the DG method to the 3D elastodynamic velocity-stress first-order system in the time domain
for an isotropic medium in a conservative form,
X
X
−
→
ρ∂t~v =
∂α (Mα~σ ) +
Mα ψα (~σ ) + f~0
α∈{x,y,z}
Λ0 ∂t~σ =
X
α∈{x,y,z}
X
∂α (Nα~v ) +
−
→
Nα ψα (~v ) + Λ0 ∂t σ~0 ,
(1)
α∈{x,y,z}
α∈{x,y,z}
where ~v is the particle velocity vector, ~σ the stress vector, ρ the density and Λ0 a matrix containing the
physical properties of the medium. Mα and Nα are matrices with real factors. In the DG formulation, the
system (1) is multiplied by a test function and integrated in the volume. We solve so-called weak form of
the system using tetrahedral cells and an explicit leap-frog scheme in time. The evolution of the system is
governed by the exchange of numerical fluxes between adjacent cells at each time step. Following the approach of BenJemaa et al. (2007), we adopt the centred flux scheme for its non-dissipative characteristic.
Moreover, we have introduced in the system (1) the CPML formulation developped by Komatitsch and
Martin (2007) which improves the behaviour of classical PML absorbing conditions (Berenger (1994))
at grazing incidence. The CPML does not require to split the elastodynamic equations, instead it makes
use of memory variables ψα which are updated at each time step according to the recursive expression,
ψαn = bα ψαn−1 + aα (∂α )n
∀α ∈ {x, y, z},
(2)
with n the time step and aα , bα some parameters depending on the time step and the length of the
absorbing layer. One memory variable is required for each partial derivative in the system (1) making
22 variables per degree of freedom in the CPML layer. We adopt the nodal form on the DG formulation
(Hesthaven and Warburton, 2008), assuming that each component of the stress or velocity vector is
decomposed as,
ndof
ndof
~vi =
X
j=1
vij ϕ
~ ij
~σi =
X
σij ϕ
~ ij ,
j=1
71st EAGE Conference & Exhibition — Amsterdam, The Netherlands, 8 - 11 June 2009
(3)
where i is the indice of the tetrahedral cell, ndof the number of degrees of freedom (or nodes), ϕ
~ ij the
interpolating Lagrange basis function at node j and vij and σij are the velocity and stress wavefields at
node j. For instance, in DG P0 , there is only one degree of freedom (the stress and velocity are constant
per cell) while in DG P1 , there are four degrees of freedom located at the 4 vertices of the tetrahedral cell
(the stress and velocity are linearly interpolated). Numerical experiments have shown that the DG P0
does not provide an acceptable solution in the 3D case due to its strong dependency to the mesh structure
and therefore one need to consider P1 as the lowest possible interpolation order to perform accurate
seismic modelling.
Computing aspects
In this section, we discuss the issues of performance and load balancing, a critical point in our application
based on a parallel architecture with MPI communications like most of geophysical applications. In table
1, we have reported the memory requirement and computation cost for updating the wavefield values in
a single cell during one time step. These simulations were performed on the IBM Blue Gene/P machine
with Power PC 450 CPUs of the IDRIS/CNRS. These figures are quite stable regardless the number of
MPI processes due to the high and intrinsic scalability of the DG method.
Cell type
Time per cell for one step Memory per cell
Ordinary P0 cell
1.6 µs
175 bytes
P0 cell inside CPML
2.5 µs
284 bytes
Ordinary P1 cell
6.2 µs
294 bytes
TABLE 1: memory and computation cost for updating the wavefield values in one cell
at one time step depending on the cell type
We first remark that a P0 cell inside the CPML costs about 60% more in time and memory than an
ordinary P0 cell due to the extra computation and the storage of the memory variables. We should also
note that the cost of a P1 cell is about 4 times more expensive in terms of computation time than the
cost of an ordinary P0 cell (and 2.8 times compared to P0 CPML cell) and approximately equivalent
regarding the memory to the cost of a P0 CPML cell. Likewise, the storage of the memory variables
in a P1 cell would require additionnal 264 bytes meaning a increase of 90% which should also reflect
the extra cost of the computation time. When dealing with large scale seismic modelling, it is of first
importance to keep the memory requirements and the computation time as low as possible and hence we
should seek for an efficient and inexpensive absorbing condition. The figures in table 1 indicate that a
good compromise should be met by combining the P1 scheme in the medium to get a acceptable accuracy
and the P0 scheme in the absorbing layers to reduce the whole numerical cost.
Numerical experiment in complex medium
In order to illustrate the proposed scheme, we considered a target of the SEG/EAGE overthrust model.
The dimensions of this target are 8 km x 8 km x 3.5 km (figure 1). An absorbing layer of 1 km width
has been added at the periphery of the model except on the top where a free surface condition has been
applied. We placed an explosive source in the middle of the model at 20 m depth and used a Ricker
wavelet with a mean frequency of 3 Hz. We built a regular mesh by dividing cubes into 5 tetrahedra and
set the cube size to tenth of the shortest propagated wavelength to reach the required precision. With a
maximum frequency of 9 Hz and a minimum S-wave velocity of 1300 m/s, the spatial discretisation is
15 m and the mesh contains more than 677 millions of tetrahedra. The simulations were performed on
the IDRIS/CNRS cluster with 4096 CPUs and a regular subdomain decomposition (figure 2). Table 2
gives the figures observed with DG P0 and DG P1 scheme when using in both cases a P0 CPML layer.
Order Time/cell Min mem/proc Max mem/proc Tot. mem Time step Nb step Tot. time
DG P0
2.4 µs
27.6 MB
44.9 MB
148 GB
0.668 ms
7489
49 min.
DG P1
6.0 µs
44.9 MB
46.4 MB
187 GB
0.200 ms 24965 6h 52 min.
TABLE 2: memory and computation cost for the simulations on the portion of the Overthurst model
with DG P0 and P1 interpolation in the medium and P0 CPML absorbing layers
71st EAGE Conference & Exhibition — Amsterdam, The Netherlands, 8 - 11 June 2009
1
2
3
Inline (km)
4
5
6
7
8
9
10
1
6000
5000
2
4000
3
m/s
Depth (km)
0
0
3000
4
2000
Figure 1: Section of the SEG/EAGE model (P-wave velocity) in the source plane. The explosive source
is placed at the middle of the model at 20 m below the free surface. A CPML layer of 1 km width has
been added to the model, the limits of this layer are plotted with yellow lines.
Depth (km)
0
0
1
2
3
Inline (km)
4
5
6
7
8
9
10
1
2
3
4
Figure 2: Map of the domain decomposition in the source plane, each subdomain associated to one CPU
is plotted with a different color. The 4096 CPUs allow for a regular domain decomposition of 32 x 32 x
4 subdomains. With such decomposition, some of these subdomains are totally located in the absorbing
layers (the limits of these layers are plotted with black lines).
For the first simulation with only P0 cells, we note that the memory is not well balanced between the
subdomains and the average computation time per cell corresponds to the cost of the P0 cells located in
the CPML layers. In this case, all the processes located in the medium spent about 60% of time waiting
for the processes located in the CPML layers. This is the main bottleneck of parallel computing, the
slowest process penalises all the others. On the contrary, for the second simulation with a mix of P1 and
P0 cells, we observe a good memory balance between the processes and the average computation time
per cell corresponds to the cost of the P1 cells located in the medium. In the second case, the absorbing
layers do not penalise the whole simulation. A series of snapshots extracted from the second simulation
are illustrated in figure (3).
Conclusions and perspectives
We have proposed a DG formulation with a low cost absorbing condition which is a crucial issue to
perform large scale seismic modelling. We obtain a good load balancing between the processes without
paying any effort on the subdomain decomposition. Nevertheless, the figures in table 2 still show a
relatively high numerical cost considering the size of the model. This has to be analysed and we should
mention that the regular mesh in the presented simulations was only used for test purpose and clarity
of demonstration. Obviously, the mesh needs to be tuned locally to the medium properties in order to
decrease significantly the number of cells and hence the global cost. For instance there is a factor of 3
between the lowest and highest velocities in the SEG/EAGE Overthrust model. The proposed scheme
will still pay off on unstructured meshes as the load balancing is achieved naturally. Another option for
reducing the CPU time, is to explore higher orders of interpolation to seek for the optimum numerical
scheme. For a complete analysis, a comparison of the DG formulation should be done against methods
71st EAGE Conference & Exhibition — Amsterdam, The Netherlands, 8 - 11 June 2009
2
Inline (km)
4
6
8
10
0
2
4
0
2
2
Inline (km)
4
6
8
10
0
0
2
Inline (km)
4
6
8
10
2
0
10
0
2
4
0
2
Inline (km)
4
6
2
4
1.5 second
10
0
2
Inline (km)
4
6
8
10
8
10
2
4.0 second
8
10
0
Depth (km)
8
8
2
2.5 second
Depth (km)
Inline (km)
4
6
Inline (km)
4
6
4
1.0 second
2
2
3.5 second
4
0
0
4
Depth (km)
0
Depth (km)
Depth (km)
0
2.0 second
4
Depth (km)
10
4
2
0
8
2
.5 second
0
Inline (km)
4
6
Depth (km)
0
Depth (km)
Depth (km)
0
0
2
Inline (km)
4
6
2
4
3.0 second
4.5 second
Figure 3: Series of snapshots in the source plane of the velocity component vx computed with the mix
DG P0 /P1 scheme. No spurious reflection at the model boundaries can be observed (the limits of the
CPML layers are plotted with yellow lines).
like FD or SEM in terms of performance and accuracy, knowing that the DG method will show its
efficiency in case of highly heterogeneous medium with or without complex topography.
Acknowledgments
This research was funded by the SEISCOPE consortium sponsored by BP, CGG-VERITAS, Exxon Mobil, SHELL and TOTAL. Access to the high performance computing facilities of IDRIS/CNRS computer
center provided the required computer ressources.
References
BenJemaa, M., Glinsky-Olivier, N., Cruz-Atienza, V. M., Virieux, J., and Piperno, S. [2007] Dynamic
non-planar crack rupture by a finite volume method. Geophysical Journal International 171, 271–285.
Berenger, J-P. [1994] A perfectly matched layer for absorption of electromagnetic waves. Journal of
Computational Physics 114, 185–200.
Dumbser, M., and Käser, M. [2006] An arbiratary high-order discontinuous Galerkin method for elastic waves on unstructured meshes - II. The three-dimentional isotropic case. Geophysical Journal
International 196(2), 319–336.
Hesthaven, Jan. S., and Warburton, Tim. [2008] Nodal Discontinuous Galerkin Method. Springer.
Komatitsch, D., and Martin, R. [2007] An unsplit convolutional perfectly matched layer improved at
grazing incidence for the seismic wave equation. Geophysics 72(5), SM155–SM167.
Komatitsch, D., and Tromp, J. [1999] Introduction to the spectral element method for 3D seismic wave
propagation. Geophys. J. Int. 139, 806–822.
71st EAGE Conference & Exhibition — Amsterdam, The Netherlands, 8 - 11 June 2009