Download Chapitre 4 (style : Chapitre)

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Radiosurgery wikipedia , lookup

Industrial radiography wikipedia , lookup

X-ray wikipedia , lookup

Backscatter X-ray wikipedia , lookup

Center for Radiological Research wikipedia , lookup

Nuclear medicine wikipedia , lookup

Positron emission tomography wikipedia , lookup

Medical imaging wikipedia , lookup

Fluoroscopy wikipedia , lookup

Image-guided radiation therapy wikipedia , lookup

Transcript
Chapter 1
Introduction to tomography
1.1. Introduction
Tomographic imaging systems are designed to analyze the structure and composition of objects by examining them with waves or radiation and by calculating virtual
cross sections through them. They cover all imaging techniques that permit the mapping of one or more physical parameters across one or more planes. In this book, we
are mainly interested in calculated, or computer aided tomography, in which the final
image of the spatial distribution of a parameter is computed from measurements of
the radiation that is emitted, transmitted, or reflected by the object. In combination
with the electronic measurement system, the processing of the collected information
thus plays a crucial role in the production of the final image. Tomography complements the range of imaging instruments dedicated to observation, such as the radar,
the sonar, the lidar, the echograph, and the seismograph. Currently, these instruments
are mostly used to detect or localize an object, for instance an airplane by its echo on
a radar screen, or to measure heights and thicknesses, for instance of the earth’s surface or of a geological layer. They mainly rely on depth imaging techniques, which
are described in another book of the French version of this series [GAL 02]. By contrast, tomographic systems calculate the value of the respective physical parameter at
all vertices of the grid that serves the spatial encoding of the image. An important
part of imaging systems such as cameras, camcorders, or microscopes is the sensor
that directly delivers the observed image. In tomography, the sensor performs indirect measurements of the image by detecting the radiation with which the object is
examined. These measurements are described by the radiation transport equations,
which lead to what the mathematicians call the direct problem, i.e. the measurement
Chapter written by Pierre GRANGEAT.
2
Tomography
or signal equation. To obtain the final image, appropriate algorithms are applied to
solve this equation and to reconstruct the virtual cross sections. The reconstruction
thus solves the inverse problem [OFT 99]. Tomography evidently yields the desired
image only indirectly by calculation.
1.2. Observing contrasts
A broad range of physical phenomena may be exploited to examine objects.
Electromagnetic waves, acoustic waves, or photonic radiation are used to carry the
information to the sensor. The choice of the exploited physical phenomenon and of
the associated imaging instrument depends on the desired contrast, which must assure a good discrimination between the different structures present in the objects.
This differentiation is characterized by its specificity, i.e. by its ability to discriminate between inconsequential normal structures and abnormal structures, and by its
sensitivity, i.e. its capacity for measuring the weakest possible intensity level of the
relevant abnormal structures. In medical imaging, for example, tumors are characterized by a metabolic hyperactivity, which leads notably to an increased glucose consumption. A radioactive marker such as fluorodeoxyglucose (FDG) permits detecting
this increase in the metabolism, but it is not absolutely specific, since other phenomena, like an inflammation, also entail a local increase in glucose consumption. Therefore, the physician must interpret the physical measure in the context of the results of
other clinical examinations.
It is preferable to use coherent radiation whenever possible, which is the case for
ultrasound, microwaves, laser radiation, and optical waves. Coherent radiation permits measuring not only its attenuation but also its dephasing. The latter serves associating a depth with the measured information, because the propagation time difference results in a dephasing. Each material is characterized by its attenuation coefficient and its refractive index, which describe the speed of propagation of waves in it.
In diffraction tomography, one essentially aims at reconstructing the surfaces of the
interfaces between materials with different indices. In materials with complex structure, however, the multiple interferences rapidly render the phase information unusable. Moreover, sources of coherent radiation are often more expensive, like
lasers. With X-rays, only phase contrast phenomena that are linked to the spatial
coherence of photons are currently observable, using microfocus sources or synchrotrons. This concept of spatial coherence reflects the fact that an interference phenomenon may only be observed behind two slits if they are separated by less than the
coherence length. In such a situation, a photon interferes solely with itself. Truly coherent X-ray sources, such as the X-FEL (X-ray Free Electron Laser), are only
emerging.
Introduction to tomography
3
In case of non-coherent radiation, like γ-rays, and in most cases of X-ray imaging
and optical imaging with conventional light sources, each material is mainly characterized by its attenuation and diffusion coefficients. The elastic or Rayleigh diffusion is
distinguished from the Compton diffusion, which results from inelastic photon-electron
collisions. In γ- and X-ray imaging, one tries to keep the attenuated direct radiation
only, since it propagates, in contrast to the diffused radiation, along a straight line and
thus permits obtaining a very high spatial resolution. The attenuation coefficients reflect
in first approximation the density of the traversed material. Diffusion tomography is, for
instance, considered in infrared imaging to study the blood and its degree of oxygenation. Since diffusion spreads light in all directions, the spatial resolution of such systems is limited.
After the interaction between radiation and matter, we now consider the principle of
generating the radiation. We distinguish between active and passive systems. In the
former, the source of radiation is controlled by an external generator that is activated at
the moment of the measurement. The object is explored by direct interrogation if the
generated incident radiation and the emerging measured radiation are of the same type.
This is the case in X-ray, optical, microwave, and ultrasound imaging, for example. The
obtained contrast consequently corresponds to propagation parameters of the radiation.
For these active systems with direct interrogation, we distinguish measurement systems
based on reflection or backscattering, where the emerging radiation is measured on the
side of the object the generator is placed, and measurement systems based on transmission, where the radiation is measured on the opposite side. The choice depends on
the type of radiation employed and on constraints linked to the overall dimensions of
the object. Magnetic resonance imaging (MRI) is a system with direct interrogation,
since the incident radiation and the emerging radiation are both radiofrequency waves.
However, the temporal variations of the excitation and receive signals are very different. The object may also be explored by indirect interrogation, which means that the
generated incident radiation and the emerging measured radiation are of different type.
This is the case in fluorescence imaging, where the incident wave produces fluorescent
light with a different wavelength (see Chap. 6). The observed contrast corresponds in
this case to the concentration of the emitter.
In contrast to active systems, passive systems rely on internal sources of radiation, which are inside the analyzed matter or patient. This is the case in magnetoand electro-encephalography (MEG and EEG), where the current dipoles linked to
synaptic activity are the sources. This is also the case in nuclear emission or photoluminescence tomography, where the tracer concentration in tissues or materials is
measured. The observed contrast again corresponds to the concentration of the
emitter. Factors such as attenuation and diffusion of the emitted radiation consequently lead to errors in the measurements.
4
Tomography
On these different systems, the contrasts may be natural or artificial. Artificial
contrasts correspond to the injection of contrast agents, such as agents based on iodine in X-ray imaging and gadolinium in MRI, and of specific tracers dedicated to
passive systems, such as tracers based on radioisotopes and fluorescence. Within the
new field of nanomedicine, new activatable markers are investigated, such as luminescent or fluorescent markers, which become active when the molecule is in a given
chemical environment, exploiting for instance the quenching effect, or when the
molecule is activated or dissociated by an external signal.
Finally, studying the kinetics of contrasts provides complementary information,
such as uptake and redistribution time constants of radioactive compounds. Moreover, multispectral studies permit characterizing flow velocity by the Doppler effect,
and dual-energy imaging allows decomposing matter into two reference materials
like water and bone, and thus enhancing the contrast.
In Tab. 1.1, principal tomographic imaging modalities are listed, including in
particular those addressed in this book. In the second column, entitled “contrast”, the
physical parameter visualized by each modality is stated. This parameter serves in
certain cases, as in functional imaging, the calculation of physiological or biological
parameters via a model, which describes the interaction of the tracer with the organism. In the third column, we propose a classification of these systems according to
their accuracy. We differentiate between very accurate systems, which are marked
by + and may reach a reproducibility of the basic measurement of 0.1 %, thus permitting quantitative imaging, accurate systems, which are marked by = and attain a
reproducibility in the order of 1 %, and less accurate systems, which are marked by –
and provide a reproducibility in the order of 10%, thus allowing qualitative imaging
only. Finally, in the fourth and fifth column, typical orders of magnitude are given
for the spatial and temporal resolution delivered by these systems. In view of the
considerable diversity of existing systems, they primarily allow a distinction between
systems with a high spatial resolution and systems with a high temporal resolution.
They also illustrate the discussion in Sec. 1.3. The spatial and temporal resolutions
are often conflicting parameters that depend on how the data acquisition and the
image reconstruction are configured. For example, increasing the time of exposure
improves the signal-to-noise ratio and thus the accuracy at the expense of the temporal resolution. Likewise, spatial smoothing alleviates the effect of noise and ameliorates the statistical reproducibility at the expense of the spatial resolution. Another
important parameter is the sensitivity of the imaging system. For instance, nuclear
imaging systems are very sensitive and are able to detect traces of radioisotopes, but
need very long acquisition times to reach a good signal-to-noise ratio. Therefore, the
accuracy is often limited. The existence of disturbing effects often corrupts the measurement in tomographic imaging and makes it difficult to attain an absolute quantification. One is then satisfied with a relative quantification, according to the accuracy
of the measurement.
Introduction to tomography
5
Tomographic
imaging
modality
Contrast
Accuracy
Spatial
resolution
Temporal
resolution
CT
attenuation coefficient,
density of matter
+
axial
 0.4 mm
0.1 s – 2.0 s
per set of slices
associated with
180° rotation
transaxial
 0.6 - 5 mm
X-ray tomography in radiology
attenuation coefficient,
density of matter
=
 0.25 mm
3-6s
by volumetric
acquisition
(180° rotation)
MRI
amplitude of transverse
magnetization
=
 0.5 mm
0.1 – 10.0 s
SPECT
tracer concentration,
associated physiological or biological parameters
-
10 - 20 mm
600 - 3 600 s
per acquisition
PET
tracer concentration,
associated physiological or biological parameters
=
3 - 10 mm
300 - 600 s
per acquisition
fMRI
amplitude of transverse
magnetization, concentration of oxy-deoxyhemoglobine
=
 0.5 mm
0.1 s
MEG
current density
-
some mm
10-3 s
EEG
current density
-
10 - 30 mm
10-3 s
TEM
attenuation coefficient
-
10-6 mm
some 100 s
(function of
image size)
CLSM
concentration of
fluorophore
-
2 · 10-4 mm
some 100 s
(function of
image size)
Coherent optical
tomography
refractive index
-
10-2 mm
10-1 s
Non-coherent
optical tomography
attenuation and diffusion coefficients
-
10 mm
some 100 s
Synchrotron
tomography
attenuation coefficient,
density of matter
+
10-3 - 10-2 mm
800 - 3 600 s
per acquisition
(180° rotation)
6
Tomography
Tomographic
imaging
modality
Contrast
Accuracy
Spatial
resolution
Temporal
resolution
X-ray tomography in NDT
attenuation coefficient,
density of matter
+
10-2 - 102 mm
(dependent on
system)
1 - 1 000 s
per acquisition
(dependent on
system)
Emission
tomography in
industrial flow
visualization
concentration of tracer
-
some 10 % of
the diameter of
the object
(some 10 mm)
10-1 s
Impedance
tomography
electric impedance
(resistance, capacitance, inductance)
-
3 - 10 % of the
diameter of the
object
(1 - 100 mm)
10-3 - 10-2 s
Microwave
tomography
refractive index
-
 - /10
with wavelength
10-2 - 10 s

(1 - 10 mm)
Industrial
ultrasound
tomography
ultrasonic refractive
index, reflective index,
speed of propagation,
attenuation coefficient
-

with wavelength
10-2 - 10-1 s

(0.4 – 10.0 mm)
+ : very accurate systems, reproducibility of the basic measurement in the order of 0.1%, quantitative imaging.
= : accurate systems, reproducibility of the basic measurement in the order of 1%.
- : less accurate systems, reproducibility of the basic measurement in the order of 10%, qualitative
imaging.
Table 1.1. Comparison of principal tomographic imaging modalities
In most cases, the interaction between radiation and matter is accompanied by
energy deposit, which may be associated with diverse phenomena, like a local increase of the thermal agitation, a change in state, an ionization of atoms, and a
breaking of chemical bonds. Improving image quality in terms of contrast- or signalto-noise ratio unavoidably leads to an increase in the applied dose. In medical imaging, a compromise has to be found to assure that image quality is compatible with
the demands of the physicians and the dose tolerated by the patient. After stopping
the irradiation, the storage of energy may be emptied by returning to the equilibrium,
by thermal dissipation, or by biological mechanisms that try to repair or replace the
defective elements.
Introduction to tomography
7
1.3. Localization in space and time
Tomographic systems provide images, i.e. a set of samples of a quantity on a
spatial grid. When this grid is two-dimensional (2D) and associated with the plane of
a cross section, one speaks of 2D imaging, where each sample represents a pixel.
When the grid is three-dimensional (3D) and associated with a volume of interest,
one speaks of 3D imaging, where each element of the volume is called a voxel.
When the measurement system provides a single image at a given moment in time,
the imaging system is static. One may draw an analogy with photography, in which
an image is instantaneously captured. When the measurement system acquires an
image sequence over a period of time, the imaging system is dynamic. One may draw
an analogy with cinematography, in which image sequences of animated scenes are
recorded. In each of these cases, the basic information associated with each sample
of the image is the local value in space and time of a physical parameter characteristic of the employed radiation. The spatial grid serves as support of the mapping
of the physical parameter. The spatial sampling step is defined by the spatial resolution of the imaging system and its ability to separate two elementary objects. The
acquisition time, which is equivalent to the time of exposure in photography, determines the temporal resolution. Each tomographic imaging modality possesses its
own characteristic spatial and temporal resolution.
When exploring an object non-destructively, it is impossible to access the desired,
local information directly. Therefore, one has to use penetrating radiation to measure
the characteristic state of the matter. The major difficulty is that the emerging radiation
will provide a global, integral measure of contributions from all traversed regions. The
emerging radiation thus delivers a projection of the desired image, in which the contribution of each point of the explored object is weighted by an integration kernel, whose
support geometrically characterizes the traversed regions. To illustrate these projective
measurements, let us consider X-ray imaging as an example, where all observed structures superimpose on a planar image. On an X-ray image of the thorax seen from the
front, the anterior and posterior parts superimpose, the diaphragm is combined with the
vertebrae. On these projection images, only the objects with high contrast are observable. In addition, only their profile perpendicular to the direction of observation may be
identified. Their 3D form may not be recognized.
To obtain a tomographic image, a set of projections is acquired under varying
measurement conditions. These variations change the integration kernel, in particular
the position of its support, to completely describe a projective transformation that is
characteristic of the employed physical means of exploration. In confocal microscopy, for example, a Fredholm integral transform characterizes the impulse response of the microscope. In MRI, it is a Fourier transform, and in X-ray tomography, it is a Radon transform. In the last case, describing the Radon transform
implies acquiring projections, i.e. radiographs, with repeated measurements under
8
Tomography
angles that are regularly distributed around the patient or the explored object. Once
the transformation of the image is described, the image may be reconstructed from
the measurements by applying the inverse transformation, which corresponds to
solving the signal equation. By these reconstruction operations, global projection
measurements are transformed into local values of the examined quantity. This localization has the effect of increasing the contrast. It thus becomes possible to discern
organs, or defects, with low contrast. This is typically the case in X-ray imaging,
where one may basically observe the bones on the radiographs, while one may identify the soft organs on the tomographs. This increase in contrast permits being more
sensitive in the search for anomalies characterized by either hyper- or hypo-attenuation, or by hyper- or hypo-uptake of a tracer. Thanks to the localization, it also becomes possible to separate organs from their environment and thus to gain characteristic information on them, such as the form of the contour, the volume, the mass,
and the number, and to position them with respect to other structures in the image.
This localization of information is of primary interest and justifies the success of
tomographic techniques. A presentation of the principles of computerized tomographic imaging can be found in [KAK 01, HSI 03, KAL 06] for X-rays and [BEN
98, BAI 05] for positron emission tomography (PET).
The inherent cost of tomographic procedures is linked to the necessity of completely describing the space of projection measurements to reconstruct the image.
Tomographic systems are thus “scanners”, i.e. systems that measure, or scan, each
point in the transformed domain. The number of basic measurements is of the same
order of magnitude as the number of vertices of the grid on which the image is calculated, typically between 642 and 10242 in 2D and between 643 and 10243 in 3D. The
technique used for data acquisition thus greatly influences the temporal resolution
and the cost of the tomographic system. The higher the number of detectors and generators placed around the object gets, the better the temporal resolution will be, but
the higher the price will rise. A major evolution in the detection of nuclear radiation
is the replacement of one-dimensional (1D) by 2D, either multi-line or surface, detectors. In this way, the measurements may be performed in parallel. In PET, one
observes a linear increase in the number of detectors available on high-end systems
with the year of the product launch. However, scanning technology must also be
taken into account for changing the conditions of the acquisition.
The most rapid techniques are those of electronic scanning. This is the case in
electromagnetic measurements and in MRI [HAA 99]. In the latter, only a single detector, a radiofrequency antenna, is used and the information is gathered by modifying the ambient magnetic fields, in such a way that the whole measurement domain is
covered. The obtained information is intrinsically 3D. The temporal resolution
achieved with these tomographic systems is very good, permitting dynamic imaging
at video rates, i.e. at about 10 images/s. When electronic scanning is impossible,
mechanical scanning is employed instead. This is the case for rotating gantries in X-
Introduction to tomography
9
ray tomography, which rotate the source-detector combination around the patient. To
improve the temporal resolution, the rotation time must be decreased - the most
powerful systems currently complete a rotation in the order of 0.3 s - and the rotation
must be combined with an axial translation of the bed to efficiently cover the whole
volume of interest. Today, it is possible to acquire the whole thoracic area in less
than 10 s, i.e. in a time that patients can hold their breath.
The principle of acquiring data by scanning, adopted by tomographic systems,
poses problems of motion compensation when objects that evolve over time are observed. This is the case in medical imaging when the patient moves or when organs
which are animated by continuous physiological motion like the heart or the lungs
are studied. This is also the case in non-destructive testing on assembly or luggage
inspection lines, or in process monitoring, where chemical or physical reaction are
followed. The problem must be considered as a reconstruction problem in space and
time, and appropriate reconstruction techniques for compensating the motion or the
temporal variation of the measured quantities must be introduced.
The evolution in tomographic techniques manifests itself mainly in the localization
in space and time. Growing acquisition rates permit increasing the dimensions of the
explored regions. Thus, the transition from 2D to 3D imaging has been accomplished.
In medical imaging, certain applications like X-ray and MR angiography and oncological imaging by PET require a reconstruction of the whole body. Accelerating the
acquisition also makes possible dynamic imaging, for instance to study the kinetics of
organs and metabolism or to guide interventions based on images. An interactive tomographic imaging is thus approached. Finally, the general trend towards miniaturization
leads to growing interest in microtomography, where an improved spatial resolution of
the systems is the primary goal. In this way, one may test micro-systems in non-destructive testing (NDT), study animal models like mice in biotechnology, or analyze
biopsies in medicine. Nanotomography of molecules such as proteins, molecule assemblies, or atomic layers within integrated microelectronic devices is also investigated
using either electron microscopy or synchrotron X-ray sources.
1.4. Image reconstruction
Tomographic systems combine the examination of matter by radiation with the calculation of characteristic parameters, which are either linked to the generation of this
radiation or the interaction between this radiation and matter. Generally, each measurement of the emerging radiation provides an analysis of the examined matter, like a
sensor. Tomographic systems thus acquire a set of partial measurements by scanning.
The image reconstruction combines these measurements to assign a local value, which
is characteristic of the employed radiation, to each vertex of the grid that supports the
spatial encoding of the image. For each measurement, the emerging radiation is the sum
of the contributions of the quantity to be reconstructed along the traversed path, each
10
Tomography
elementary contribution being either an elementary source of radiation or an elementary
interaction between radiation and matter. It is thus an integral measurement which
defines the signal equation. The set of measurements constitutes the direct problem,
linking the unknown quantities to the measurements. The image reconstruction amounts
to solving this system of equations, i.e. the inverse problem [OFT 99].
The direct problem is deduced from the fundamental equations of the underlying
physics, such as the Boltzmann equation for the propagation of photons in matter, the
Maxwell equation for the propagation of electromagnetic waves [DEH 95], and the
elastodynamic equation for the propagation of acoustic waves [DEH 95]. The measurements provide integral equations, often entailing numerically instable differentiations in
the image reconstruction. In this case, the reconstruction corresponds to an inverse
problem that is ill-posed in the sense of Hadamard [HAD 32].
The acquisition is classified as complete when it provides all measurements that are
necessary to calculate the object, which is in particular the case when the measurements
cover the whole support of the transformation modeling the system. The inverse problem is considered weakly ill-posed, if the principal causes of errors are instabilities
associated with noise or the bias introduced by an approximate direct model. This is the
case in X-ray tomography, for instance. If additionally different solutions may satisfy
the signal equation, even in absence of noise, the inverse problem is considered strongly ill-posed. This is the case in magneto-encephalography and in X-ray tomography
using either a small angular range or a small number of positions of the X-ray source,
for example.
The acquisition is classified as incomplete when some of the necessary measurements are missing, in particular when technological constraints reduce the number of
measurements that can be made. This situation may be linked to a subsampling of
measurements or an absence of data caused by missing or truncated acquisitions. In this
case, the inverse problem is also strongly ill-posed, and the reconstruction algorithm
must choose between several possible solutions.
To resolve these ill-posed inverse problems in the sense of Hadamard, the reconstruction must be regularized by imposing constraints on the solution. The principal
techniques of regularization reduce the number or the range of the unknown parameters – for instance by using a coarser spatial grid to encode the image, by parameterizing the unknown image, or by imposing models on the dynamics –, introduce
regularity constraints on the image to be reconstructed, or eliminate from the inverse
operator the smallest spectral or singular values of the direct operator. All regularization techniques require that a compromise is made between the fidelity to the
given data and the fidelity to the constraints on the solution. This compromise reflects
the uncertainty relations between the localization and the quantification in images, applied to tomographic systems.
Introduction to tomography
11
Several approaches may be pursued to define algorithms for image reconstruction,
for which we refer the reader to [HER 79, NAT 86, NAT 01]. Like for numerical
methods in signal processing, deterministic and statistical methods, and continuous and
discrete methods are distinguished. For a review on recent advances in image reconstruction, we refer the reader to [CEN 08].
The methods most widely used are continuous approaches, also called analytical
methods. The direct problem is described by an operator over a set of functions. In the
case of systems of linear measurements, the unknown image is linked to the measurements by a Fredholm transform of the first kind. In numerous cases, this equation simplifies either to a convolution equation, a Fourier transform, or a Radon transform. In
these cases, the inverse problem admits an explicit solution, either in form of an inversion formula, or in form of a concatenation of explicit transformations. A direct method
of calculating the unknown image is obtained in this way. These methods are thus more
rapid than the iterative methods employed for discrete approaches. In addition, these
methods are of general use, because the applied regularization techniques, which perform a smoothing, do not rely on specific a priori knowledge about the solution. The
discretization is only introduced for the numerical implementation.
However, when the direct operator becomes more complex to better describe the
interaction between radiation and matter, or when the incorporation of a priori information is necessary to choose a solution in case of incomplete data, an explicit formula
for calculating the solution does not exist anymore. In these specific cases, discrete approaches are preferable. They describe the measurements and the unknown image as
two finite, discrete sets of sampled values, represented by vectors. When the measurement system is linear or can be linearized by a simple transformation of the data, the
relation between the image and the measurements may be expressed by a matrix-vector
product. The problem of image reconstruction then leads to the problem of solving a
large system of linear equations. In the case of slightly underdetermined, complete
acquisitions, regularization is introduced by imposing smoothness constraints. The
solution is formulated as the argument of a composite criterion that combines the
fidelity to the given data and the regularity of the function. In the case of highly underdetermined, incomplete acquisitions, regularization is introduced by choosing an optimal solution among a set of possible solutions, which are, for example, associated with
a discrete subspace that depends on a finite number of parameters. The solution is thus
defined as the argument of a linear system under constraints. In both cases, the solution
is calculated with iterative algorithms.
When the statistic fluctuation of the measurements becomes important, as in
emission tomography, it is preferable to explicitly describe the images and measurements statistically, following the principle of statistical methods. The solution is expressed as a statistical estimation problem. In the case of weakly ill-posed problems,
Bayesian estimators corresponding to the minimization of a composite criterion have
12
Tomography
been proposed. They choose the solution with the highest probability for the processed measurement, or on average for all measurements, taking the measurements and
the a priori information about the object into account. In the case of strongly illposed problems, constrained estimators, for example linked to Kullback distances,
have been envisioned. Also in these cases, one resorts to iterative reconstruction algorithms.
When the ill-posed character becomes too severe, for example in case of very
limited numbers of measurement angles, constraints at low levels of the image samples are no longer sufficient. One must then employ a geometric description of the
object to be reconstructed, in form of a set of elementary objects, specified by a
graphical model or a set of primitives. The image reconstruction amounts to matching the calculated and the measured projections. In this work, these approaches,
which do not directly concern the problem of mapping a characteristic physical parameter, but rather problems in computer vision and pattern recognition, are not considered any further. We refer the reader to [GAR 06].
The image reconstruction algorithms require considerable processing performance. The elementary operations are the backprojection, or the backpropagation, in
which a basic measurement is propagated through a volume by following the path
traversed in the acquisition in reverse direction. For this reason, the use of more and
more sophisticated algorithms, and the reconstruction of larger and larger images,
with higher spatial and temporal resolution, are only made possible by advances in
the processing performance and the memory size of the employed computers. Often,
the use of dedicated processors and of parallelization and vectorization techniques is
necessary. In view of these demands, dynamic tomographic imaging in real-time
constitutes a major technological challenge today.
1.5. Application domains
By using radiation to explore the matter and calculations to compute the map of the
physical parameter characteristic of the employed radiation, tomographic imaging
systems provide virtual cross sections through objects without resorting to destructive
means. These virtual cross sections may easily be interpreted and enable the identification and localization of internal structures. Computed tomography techniques directly
provide a digital image, thus offering the possibility of applying the richness of image
processing algorithms. The examination of these images notably permits a characterization of each image element for differentiation, a detection of defects for control and
testing, a quantification for measurement, and a modeling for description and understanding. The imaging becomes 3D when the tomographic system provides a set of
parallel cross sections that cover the region of interest. Thus, cross sections in arbitrary
planes may be calculated, and surfaces, shapes, and the architecture of internal struc-
Introduction to tomography
13
tures may be represented. In applications associated with patients or materials with
solid structures in which fluids circulate, a distinction is made between imaging systems
that primarily map the solid architecture, i.e. the structures – one then speaks of morphological imaging – and imaging systems that mainly follow the exchange of fluids,
i.e. the contents – one then speaks of functional imaging.
In optical imaging, the development of the first, simple microscopes, which were
equipped with a single lens, dates back to the 17th century. From a physical point of
view, the use of X-rays for the examination of matter started with their discovery by
Röntgen in 1895. In the course of experiments conducted to ascertain the origin of this
radiation, Becquerel found the radiation that uranium spontaneously emits in 1896
[BIM 99]. After the discovery of polonium and radium in 1898, Pierre and Marie Curie
characterized the phenomenon that produces this radiation and called it “radioactivity”.
For these discoveries, Henri Becquerel, Marie Curie and Pierre Curie received the
Nobel price for physics in 1903. The use of radioactive tracers was initiated in 1913 by
Hevesy, Nobel price for chemistry in 1943. The discovery of artificial radioactivity by
Irène and Frédéric Joliot-Curie in 1934 opened the way for the use of numerous tracers
to study the functioning of the living, from cells to entire organisms. Important research
has then been carried out on the production and the detection of these radiations. The
first microscope using electrons to study the structure of an object was developed by
Ernst Ruska in 1929. From a mathematical point of view, the Fourier transform
constitutes an important tool in the study of the analytical methods used in tomography.
Several scholars like Maxwell, Helmholtz, and Born have studied the equation of wave
propagation. The name of Fredholm remains linked to the theory of integral equations.
The problem of reconstructing an image from line integrals goes back to the original
publication by Radon in 1917 [RAD 17], reissued in 1986 [RAD 86], and the studies of
John published in 1934 [JOH 34].
During the second half of the 20th century, a rapid expansion of tomographic techniques took place. In X-ray imaging, the first tomographic systems were mechanic
devices which combined movements of the source and of the radiological film to eliminate information off the plane of the cross section by a blurring effect. This is called
longitudinal tomography, since the plane of the cross section is parallel to the patient
axis. The first digitally reconstructed images were obtained by radio astronomers at the
end of the 1950s [BRA 79]. The first studies of Cormack at the end of the 1950s and
the beginning of the 1960s showed that it is possible to precisely reconstruct the image
of a transverse cross section of a patient from its radiographic projections [COR 63,
COR 64]. In the 1960s, several experiments with medical applications were carried out
[COR 63, COR 64, KUH 63, OLD 61]. The problem of reconstructing images from
projections in electron microscopy was also tackled in this time [DER 68, DER 71]. In
1971, Hounsfield built the first prototype of an X-ray tomograph dedicated to cerebral
imaging [HOU 73]. Hounsfield and Cormack received the Nobel price for physiology
and medicine in 1979 [COR 80, HOU 80]. The beginnings of MRI go back to the first
14
Tomography
publication by Lauterbur in 1973 [LAU 73]. In nuclear medicine, the pioneering work
in single photon emission computed tomography (SPECT) date back to Kuhl and Edwards in 1963 [KUH 63], followed by Anger in 1967 [ANG 67], Muehllehner in 1971
[MUE 71], and Keyes et al. [BOW 73, KEY 73]. The first PET scanner was built at the
beginning of the 1960s by Rankowitz et al. [RAN 62], followed in 1975 by TerPogossian and Phelps [PHE 75, TER 75]. For more references covering the beginning
of tomography, we refer the interested reader to the works of Herman [HER 79], Robb
[ROB 85], and Webb [WEB 90], and to the following papers summarizing the developments achieved during the past 50 years in X-ray computed tomography [KAL 06],
MRI [MAL 06], SPECT [JAS 06], PET [MUE 06], and image reconstruction [DEF
06].
In France, we keep of this time the pioneering work of the group of Robert Di
Paola at the Institute Gustave Roussy (IGR) on gamma tomography and of the
groups of the Laboratory of Electronics, Technology, and Instrumentation (LETI) at
the Atomic Energy Commission (CEA). In 1976, under the direction of Edmond
Tournier and Robert Allemand, the first French prototype of an X-ray tomograph
was installed at the Grenoble hospital, in collaboration with the company CGR [PLA
98]. In 1981, the LETI installed at the Service Hospitalier Frédéric Joliot (SHFJ) of
the CEA its first prototype of a time-of-flight PET scanner (TTV01), followed by
three systems of the next generation (TTV03). The SHFJ participated in the specification of these systems and carried out their experimental evaluation. The SHFJ was
the first center in France having a cyclotron for medical use and a radiochemistry
laboratory for synthesizing the tracers used in PET. In 1990, two prototypes of 3D
X-ray scanners called morphometers were developed by the LETI, one of which was
installed at the hospital of Rennes and the other at the neurocardiological hospital of
Lyon, in collaboration with the company General Electric-MSE and the groups of
these hospitals. In parallel, several prototypes of X-ray scanners were built for applications in non-destructive testing.
The 1980s saw the development of 3D imaging and digital detection technology, as
well as tremendous advances in the processing performance of computers. One pioneer
project was the DSR project (Dynamic Spatial Reconstructor at the Mayo Clinic in
Rochester [ROB 85], which aimed at imaging the beating heart and the breathing lung
using several pairs of X-ray sources and detectors. The evolution continued in the
1990s, focusing notably on increasing the sensitivity and the acquisition rates of the
systems, the latter in particular linked with the development of 2D multi-line or surface
radiation detectors. Several references describe the advances achieved in this period
[BEN 98, GOL 95, KAL 00, LIA 00, ROB 85, SHA 89].
Medical imaging is an important application domain for tomographic techniques.
Currently, X-, -, and -ray tomography, and MRI together make up a third of the
medical imaging market. Tomographic systems are heavy and expensive devices that
Introduction to tomography
15
often require specific infrastructure, like protection against ionizing radiation or
magnetic fields or facilities for preparing radiotracers. Nevertheless, tomography
plays a crucial role in diagnostic imaging today, and it has a major impact on public
health.
Medical imaging and the life sciences thus constitute the primary application domains of tomographic systems. In the life sciences, the scanning transmission electron
microscope permits studying fine structures, like those of proteins or viruses, and, at a
larger scale, the confocal laser scanning microscope (CLSM) allows visualizing the
elements of cells. In neuroscience, PET, fMRI (functional MRI), MEG, and EEG have
contributed to the identification and the localization of cortical areas and thus to the
understanding of the functioning of the brain. By permitting a tracking of physiological
parameters, PET contributes to the study of primary and secondary effects of drugs.
Finally, the technique of gene reporters opens up the possibility of genetic studies in
vivo. For these studies on drugs or in biotechnology, animal models are necessary,
which explains the current development of platforms for tomographic imaging on
animals. For small animal studies, optical tomography associated either with luminescence or laser-induced fluorescence is also progressing. In particular, it allows linking
in vitro studies on cell cultures and in vivo studies on small animal using the same
markers. Generally speaking, through MRI, nuclear and optical imaging, molecular
imaging is an active field of research nowadays.
In medicine, X-ray tomography and MRI are anatomic, morphological imaging
modalities, while SPECT, PET, fMRI, and magneto- and electroencephalography
(MEG and EEG) are employed in functional studies. X-ray tomography is well suited
for examining dense structures, like the bones or the vasculature after injection of a
contrast agent. MRI is better suited for visualizing contrast based on the density of
hydrogen atoms, as between the white and gray matter in the brain [HAA 99]. SPECT
is used a lot to study the blood perfusion of the myocardium, and PET is about to establish itself for staging of cancer in oncology. For medical application, optical tomography [MÜL 93] remains for the moment restricted to a few dedicated applications,
such as the imaging of the cornea or the skin by optical coherence tomography. Other
applications like the exploration of the breast or prostate using laser-induced fluorescence tomography are still under investigation. These aspects are elaborated in the following chapters that are dedicated to the individual modalities.
Tomographic system are used in all stages of patient care, including the diagnosis,
the therapy or surgical planning, the image-based guidance of interventions for minimal invasive surgery, and the therapy follow-up. Digital assistants may help with the
image analysis. Moreover, robots use these images for guidance. These digital images
are easily transmitted over imaging networks and stored to establish a reference database, to be shared among operators or students. In the future, veritable virtual patients
16
Tomography
and digital anatomic atlases will serve the training of certain medical operations, or
even surgical procedures. Such digital models also facilitate the design of prostheses.
The second application domain of tomographic imaging systems are astronomy
and the material sciences. They often involve large instruments, such as telescopes,
synchrotrons, and plasma chambers, and nuclear simulation devices, such as the installations Mégajoule and Airix in France. Geophysics is another important domain in
which, on a large scale, oceans or geological layers are examined and, on a smaller
scale, rocks and their properties, such as their porosity. Astrophysics, associated with
telescopes operating at different ranges of wavelengths, is also concerned, which
exploits the natural rotation of the stars for tomographic acquisitions [BRA 79, FRA
05].
The last application domain is the industry and services. Several industries are
interested, like the nuclear, aeronautic, automobile, agriculture, food, pharmaceutical,
petroleum, and chemical industry [BAR 00, BEC 97, MAY 94]. Here, a broad range of
materials in different phases is concerned. Tomography is first of all applied in design
and development departments, for the development of fabrication processes, the characterization and modeling of the behavior of materials, reverse engineering, and the
adjustment of graphical models in computer aided design (CAD). Second, tomography
is also applied in process control [BEC 97], for instance in mixers, thermal exchangers,
multi-phase transport, fluidified bed columns, motors, and reactors. It concerns in
particular the identification of functioning regimes, the determination of resting times
and of the kinetics of products, and the validation of thermo-hydraulic models. However, in non-destructive testing, once the types of defects or malfunction are identified
on a prototype, one often tries to reduce the number of measurements and thus the cost,
and to simplify the systems. Therefore, the use of tomographic systems remains limited
in this application domain. Lastly, tomography is applied to security and quality
control. This includes for instance production control, such as on-line testing after assembly to check dimensions or to study defects, or the inspection of nuclear waste
drums. Security control is also becoming a major market. This concerns in particular
luggage inspection with identification of materials to detect drugs or explosives.
In conclusion, from the mid of the 20th century up to now, tomography has become
a major technology with a large range of applications.
1.6. Bibliography
[ANG 67] ANGER H, PRICE DC, YOST PE, “Transverse section tomography with the gamma
camera”, J. Nucl. Med., vol. 8, pp. 314-315, 1967.
[BAI 05] BAILEY DL, TOWNSEND DW, VALK PE, MAISEY MN, Positron Emission Tomography: Basic Science and Clinical Practice, Springer, 2005.
Introduction to tomography
17
[BAR 00] BARUCHEL J, BUFFIÈRE JY, MAIRE E, MERLE P, PEIX G (Eds.), X-Ray Tomography
in Material Science, Hermès, 2000.
[BEC 97] BECK MS, DYAKOWSKI T, WILLIAMS RA, “Process tomography – the state of the
art”, Frontiers in Industrial Process Tomography II, Proc. Engineering Foundation Conference, pp. 357-362, 1997.
[BEN 98] BENDRIEM B, TOWNSEND D (Eds.), The Theory and Practice of 3D PET, Kluwer
Academic Publishers, Developments in Nuclear Medicine Series, vol. 32, 1998.
[BIM 99] BIMBOT R, BONNIN A, DELOCHE R, LAPEYRE C, Cent Ans après, la Radioactivité, le
Rayonnement d’une Découverte, EDP Sciences, 1999.
[BOW 73] BOWLEY AR, TAYLOR CG, CAUSER DA, BARBER DC, KEYES WI, UNDRILL PE,
CORFIELD JR, MALLARD JR, “A radioisotope scanner for rectilinear arc, transverse section,
and longitudinal section scanning (ASS, the Aberdeen Section Scanner)”, Br. J. Radiol.,
vol. 46, pp. 262-271, 1973.
[BRA 79] BRACEWELL RN, “Image reconstruction in radio astronomy”, in Herman GT (Ed.),
Image Reconstruction from Projections, Springer, pp. 81-104, 1979.
[CEN 08] CENSOR Y, JIANG M, LOUIS AK (Eds.), Mathematical Methods in Biomedical Imaging and Intensity-Modulated Radiation Therapy (IMRT), Edizioni Della Normale,
2008.
[COR 63] CORMACK AM, “Representation of a function by its line integrals, with some radiological applications”, J. Appl. Phys., vol. 34, pp. 2722-2727, 1963.
[COR 64] CORMACK AM, “Representation of a function by its line integrals, with some radiological applications II”, J. Appl. Phys., vol. 35, pp. 195-207, 1964.
[COR 80] CORMACK AM, “Early two-dimensional reconstruction (CT scanning) and recent
topics stemming from it, Nobel Lecture, December 8, 1979”, J. Comput. Assist. Tomogr.,
vol. 4, n° 5, p. 658, 1980.
[DEF 06] DFRISE M, GULLBERG GT, “Image reconstruction”, Phys. Med. Biol., vol. 51, n° 13,
pp. R139-R154, 2006.
[DEH 95] DE HOOP AT, Handbook of Radiation and Scattering of Waves, Academic Press,
1995.
[DER 68] DE ROSIER DJ, KLUG A, “Reconstruction of three-dimensional structures from
electron micrographs”, Nature, vol. 217, pp. 130-134, 1968.
[DER 71] DE ROSIER DJ, “The reconstruction of three-dimensional images from electron
micrographs”, Contemp. Phys., vol. 12, pp. 437-452, 1971.
[FRA 05] FRAZIN RA, KAMALABADI F, “Rotational tomography for 3D reconstruction of the
white-light and EUV corona in the post-SOHO era”, Solar Physics, vol. 228, pp. 219237, 2005.
[GAL 02] GALLICE J (Ed.), Images de Profondeur, Hermès, 2002.
[GAR 06] GARDNER RJ, Geometric Tomography, Encyclopedia of Mathematics and its Applications, Cambridge University Press, 2006.
18
Tomography
[GOL 95] GOLDMAN LW, FOWLKES JB (Eds.), Medical CT and Ultrasound: Current Technology and Applications, Advanced Medical Publishing, 1995.
[HAA 99] HAACKE E, BROWN RW, THOMPSON MR, VENKATESAN R, Magnetic Resonance
Imaging. Physical Principles and Sequence Design, John Wiley and Sons, 1999.
[HAD 32] HADAMARD J, Le Problème de Cauchy et les Equations aux Dérivées Partielles
Linéaires Hyperboliques, Hermann, 1932.
[HER 79] HERMAN GT (Ed.), Image Reconstruction from Projections: Implementation and
Applications, Springer, 1979.
[HOU 73] HOUNSFIELD GN, “Computerized transverse axial scanning (tomography). I. Description of system”, Br. J. Radiol., vol. 46, pp. 1016-1022, 1973.
[HOU 80] HOUNSFIELD GN, “Computed medical imaging, Nobel Lecture, December 8,
1979”, J. Comput. Assist. Tomogr., vol. 4, n° 5, p. 665, 1980.
[HSI 03] HSIEH J, Computed Tomography: Principles, Design, Artifacts, and Recent Advances, SPIE Press, 2003.
[JAS 06] KASZCZAK RJ, “The early years of single photon emission computed tomography
(SPECT): an anthology of selected reminiscences”, Phys. Med. Biol., vol. 51, n°13,
pp. R99-R115, 2006.
[JOH 34] JOHN F, “Bestimmung einer Funktion aus ihren Integralen über gewisse Mannigfaltigkeiten”, Mathematische Annalen, vol. 109, pp. 488-520, 1934.
[KAL 00] KALENDER WA, Computed Tomography, Wiley, New York, 2000.
[KAL 06] KALENDER WA, Computed Tomography: Fundamentals, System Technology,
Image Quality, Applications, Wiley-VCH, 2006.
[KAK 01] KAK AC, SLANEY M, Principles of Computerized Tomographic Imaging, SIAM,
2001.
[KEY 73] KEYES JW, KAY DB, SIMON W, “Digital reconstruction of three-dimensional radionuclide images”, J. Nucl. Med., vol. 14, pp. 628-629, 1973.
[KUH 63] KUHL DE, EDWARDS RQ, “Image separation radioisotope scanning”, Radiology,
vol. 80, pp. 653-661, 1963.
[LAU 73] LAUTERBUR PC, “Image formation by induced local interactions: examples employing nuclear magnetic resonance”, Nature, vol. 242, pp. 190-191, 1973.
[LIA 00] LIANG ZP, LAUTERBUR PC, Principles of Magnetic Resonance Imaging. A Signal
Processing Perspective, IEEE Press, Series in Biomedical Imaging, 2000.
[MAL 06] MALLARD JR, “Magnetic resonance imaging - the Aberdeen perspective on developments in the early years”, Phys. Med. Biol., vol. 51, n° 13, pp. R45-R60, 2006.
[MAY 94] MAYINGER F (Ed.), Optical Measurements: Techniques and Applications, Springer, 1994.
[MUE 71] MUEHLLEHNER G, WETZEL R, “Section imaging by computer calculation”, J. Nucl.
Med., vol. 12, pp. 76-84, 1971.
Introduction to tomography
19
[MUE 06] MUEHLLEHNER G, KARP JS, “Positron emission tomography”, Phys. Med. Biol.,
vol. 51, n° 13, pp. R117-R137, 2006.
[MÜL 93] MÜLLER G, CHANCE B, ALFANO R, et al. Medical Optical Tomography: Functional
Imaging and Monitoring, SPIE Optical Engineering Press, vol. IS11, 1993.
[NAT 86] NATTERER F, The Mathematics of Computerized Tomography, John Wiley and
Sons, 1986.
[NAT 01] NATTERER F, WUBBELING F, Mathematical Methods in Image Reconstruction,
SIAM, 2001.
[OFT 99] OBSERVATOIRE FRANÇAIS DES TECHNIQUES AVANCEES, Problèmes Inverses : de
l’Expérimentation à la Modélisation, Tec & Doc, Série ARAGO, n° 22, 1999.
[OLD 61] OLDENDORF WH, “Isolated flying detection of radiodensity discontinuities displaying the internal structural pattern of a complex object”, IRE Trans. BME, vol. 8, pp. 6872, 1961.
[PHE 75] PHELPS ME, HOFFMAN EJ, MULLANI NA, TER-POGOSSIAN MM, “Application of annihilation coincidence detection to transaxial reconstruction tomography”, J. Nucl. Med.,
vol. 16, pp. 210-224, 1975.
[PLA 98] PLAYOUST B, De l’Atome à la Puce – Le LETI: Trente Ans de Collaborations
Recherche Industrie, LIBRIS, 1998.
[RAD 17] RADON J, “Über die Bestimmung von Funktionen durch ihre Integralwerte längs
gewisser Mannigfaltigkeiten”, Math.-Phys. Kl. Berichte d. Sächsischen Akademie der
Wissenschaften, vol. 69, pp. 262-267, 1917.
[RAD 86] RADON J, “On the determination of functions from their integral values along certain manifolds”, IEEE Trans. Med. Imaging, pp. 170-176, 1986.
[RAN 62] RANKOWITZ S, “Positron scanner for locating brain tumors”, IEEE Trans. Nucl.
Sci., vol. 9, pp. 45-49, 1962.
[ROB 85] ROBB RA (Ed.), Three-dimensional Biomedical Imaging, vol. I and II, CRC Press,
1985.
[SHA 89] SHARP PF, GEMMEL HG, SMITH FW (Eds.), Practical Nuclear Medicine, IRL Press
at Oxford University Press, Oxford, 1989.
[TER 75] TER-POGOSSIAN MM, PHELPS ME, HOFFMAN EJ, MULLANI NA, “A positron
emission transaxial tomograph for nuclear imaging (PETT)”, Radiology, vol. 114, pp. 8998, 1975.
[WEB 90] WEBB S, From the Watching of the Shadows: the Origins of Radiological Tomography, Adam Hilger, 1990.
[WIL 95] WILLIAMS RA, BECK MS (Eds.), Process Tomography: Principles, Techniques and
Applications, Butterworth Heinemann, 1995.
20
Tomography
accuracy, 4
acoustic waves, 2
acquisition time, 7
active system, 3
analytical methods, 11
anatomic imaging, 15
Anger, 14
astronomy, 16
attenuation coefficient, 2, 3
backprojection, 12
backpropagation, 12
backscattering, 3
Boltzmann equation, 10
CEA, 14
coherent radiation, 2
complete, 10
computed tomography, 12
computer vision, 12
concentration, 3
continuous method, 11
contrast, 3, 4, 8
contrast agent, 4
Cormack, 13
cost, 8
density, 3
design, 16
deterministic method, 11
diagnosis, 15
diffusion coefficient, 3
direct interrogation, 3
direct problem, 1, 10
discrete method, 11
dual-energy, 4
dynamic, 7
dynamic imaging, 9, 12
Dynamic Spatial Reconstrutor, 14
Edwards, 14
elastodynamic equation, 10
electromagnetic waves, 2
electronic scanning, 8
estimation, 11
fluorescence, 3
Fourier transform, 7, 11
Fredholm transform, 7, 11
functional imaging, 13
gamma ray, 3
Hounsfield, 13
ill-posed, 10
image-based guidance, 15
incomplete, 10
indirect interrogation, 3
indirect measurements, 1
industry, 16
integral measurement, 10
inverse problem, 2, 10
iterative algorithm, 11, 12
kinetics, 4, 9
Kuhl, 14
Lauterbur, 14
LETI, 14
life science, 15
localization, 8
magnetic resonance imaging, 3
material science, 16
matrix-vector product, 11
Maxwell equation, 10
mechanical scanning, 8
medical imaging, 14, 15
microtomography, 9
minimal invasive surgery, 15
molecular imaging, 15
morphological imaging, 13, 15
morphometer, 14
motion compensation, 9
non-coherent radiation, 3
passive system, 3
pattern recognition, 12
Phelps, 14
photonic radiation, 2
pixel, 7
process control, 16
projection, 7
Introduction to tomography
public health, 15
quality control, 16
Radon, 13
Radon transform, 7, 11
Rankowitz, 14
reconstruction, 8, 9, 12
reflection, 3
refractive index, 2
regularization, 10, 11
security, 16
sensitivity, 2
service, 16
SHFJ, 14
spatial grid, 7
spatial resolution, 4, 7
specificity, 2
static, 7
21
statistical method, 11
strongly ill-posed, 10
surgical planning, 15
temporal resolution, 4, 8
Ter-Pogossian, 14
therapy, 15
therapy follow-up, 15
three-dimensional imaging, 7, 9, 14
tomographic imaging systems, 1
tomography, 1, 15
transmission, 3
two-dimensional imaging, 7, 9
uncertainty relation, 10
voxel, 7
weakly ill-posed, 10
whole body, 9
X-ray, 3, 7