Lecture Notes in Mathematics
Lecture Notes in Mathematics
net/publication/226621630
CITATIONS READS
31 9,794
7 authors, including:
All content following this page was uploaded by Maria-Luisa Rapun on 13 March 2014.
So they already know what the C.I.M.E. is all about. For the benefit of future potential users and co-
operators the main purposes and the functioning of the Centre may be summarized as follows: every
year, during the summer, Sessions (three or four as a rule) on different themes from pure and applied
mathematics are offered by application to mathematicians from all countries. Each session is generally
based on three or four main courses (24−30 hours over a period of 6-8 working days) held from
specialists of international renown, plus a certain number of seminars.
A C.I.M.E. Session, therefore, is neither a Symposium, nor just a School, but maybe a blend of both.
The aim is that of bringing to the attention of younger researchers the origins, later developments, and
perspectives of some branch of live mathematics.
The topics of the courses are generally of international resonance and the participation of the courses
cover the expertise of different countries and continents. Such combination, gave an excellent opportu-
nity to young participants to be acquainted with the most advance research in the topics of the courses
and the possibility of an interchange with the world famous specialists. The full immersion atmosphere
of the courses and the daily exchange among participants are a first building brick in the edifice of
international collaboration in mathematical research.
Inverse Problems
and Imaging
Chapters by:
A. Carpio · O. Dorn · M. Moscoso · F. Natterer
G.C. Papanicolaou · M.L. Rapún · A. Teta
ABC
Luis L. Bonilla (Ed.) George C. Papanicolaou
Gregorio Millán Institute of Fluid Dynamics Mathematics Department
Nanoscience and Industrial Mathematics Building 380, 383V
Universidad Carlos III de Madrid Stanford University
Avda. de la Universidad 30 Stanford, CA 94305, USA
28911 Leganés, Madrid, Spain [email protected]
[email protected]
http://scala.uc3m.es Maria Luisa Rapún
Departamento de Fundamentos
Ana Carpio Matemáticos de la Tecnología Aeronáutica
Departamento de Matemática Aplicada Escuela Técnica Superior de Ingenieros
Facultad de Matemáticas Aeronáuticos, Universidad Politécnica
Universidad Complutense de Madrid de Madrid, Plaza del Cardenal Cisneros 3
Plaza de Ciencias 3 28040 Madrid, Spain
28040 Madrid, Spain [email protected]
[email protected]
www.mat.ucm.es/∼acarpio Alessandro Teta
Dipartimento di Matematica Pura
Oliver Dorn e Applicata, Università di L’Aquila
Miguel Moscoso via Vetoio – loc. Coppito
Gregorio Millán Institute of Fluid Dynamics 67100 L’Aquila, Italy
Nanoscience and Industrial Mathematics [email protected]
Universidad Carlos III de Madrid
Avda. de la Universidad 30
28911 Leganés, Madrid, Spain
[email protected]
[email protected]
Frank Natterer
Institut für Numerische und Angewandte
Mathematik, University of Münster
Einsteinstraße 62
48149 Münster
Germany
[email protected]
The lectures by Prof. Frank Natterer introduce the mathematical theory and
the reconstruction algorithms of computerized X-ray tomography. These lectures
give a short account of integral geometry and the Radon transform, reconstruc-
tion algorithms such as the filtered back projection algorithm, iterative methods
(for example, the Kaczmarz method) and Fourier methods. They also comment
on the three-dimensional case, which is the subject of current research. Many of
the fundamental tools and issues of computerized tomography, such as back pro-
jection, sampling, and high frequency analysis, have their counterparts in more
advanced imaging techniques for impedance, optical or ultrasound tomography
and are most easily studied in the framework of computerized tomography.
The chapter by O. Dorn, H. Bertete-Aguirre and G.C. Papanicolaou reviews
electromagnetic induction tomography, used to solve imaging problems in geo-
physical and environmental imaging applications. The focus is on realistic 3D
situations which provide serious computational challenges as well as interesting
novel mathematical problems to the practitioners. The chapter first introduces
the reader to the mathematical formulation of the underlying inverse problem;
it then describes the theory of sensitivity analysis in this application; it proposes
a nonlinear reconstruction algorithm for solving such problems efficiently; it dis-
cusses a regularization technique for stabilizing the reconstruction; and finally it
presents various numerical examples for illustrating the discussed concepts and
ideas.
The chapter by M. Moscoso presents optical imaging of biological tissue using
the polarization effects of a narrow beam of light. The biological tissue is modeled
as a continuous medium which varies randomly in space and which contains
inhomogeneities with no sharp boundaries. This differs from the more usual
point of view in which the biological tissue is modeled as a medium containing
discrete spherical particles of the same or different sizes. The propagation of
light is then described by a vector radiative transport equation which is solved
by a Monte Carlo method. A discussion on how to use polarization to improve
image reconstruction is given as well.
The chapter by A. Carpio and M.L. Rapún explains how to use topological
derivative methods to solve constrained optimization reformulations of inverse
scattering problems. This chapter gives formulas to calculate the topological
derivatives for the Helmholtz equation and for the equations of elastic waves.
Furthermore they explain and implement a practical iterative numerical scheme
to detect objects based on computing the topological derivative of a cost func-
tional associated to these equations in successive approximate domains. Many
examples of reconstruction of objects illustrate this method.
The chapter by O. Dorn deals with an inverse problem in underwater acoustic
and wireless communication. He establishes a link between the time-reversal
and adjoint methods for imaging and proposes a method for solving the inverse
problem based on iterative time-reversal experiments.
Lastly, the chapter by G. Dell’Antonio, R. Figari and A. Teta reviews the
theory of Hamiltonians with point interactions, i.e., with potentials supported on
a finite set of points. This chapter studies the mathematical basis of scattering
Preface VII
with point scatterers, analyzes how such an idealized situation relates to short-
range potentials and discusses situations in which the strength of the potential
depends on the wave function as it has been proposed in the physics literature
on double barriers and other nanostructures. Knowing the solution of the di-
rect problem is of course a prerequisite to be able to image the scatterers from
measurements on a boundary.
X-ray Tomography
Frank Natterer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2 Integral Geometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1 The Radon Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.2 The Ray Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.3 The Cone-Beam Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2.4 The Attenuated Radon Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
2.5 Vectorial Transforms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3 Reconstruction Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.1 The Filtered Backprojection Algorithm . . . . . . . . . . . . . . . . . . . . . . . . 26
3.2 Iterative Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.3 Fourier Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4 Reconstruction from Cone Beam Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
X Contents
Miguel Moscoso
1 Introduction
Imaging is a broad field which covers all aspects of the analysis, modification,
compression, visualization, and generation of images. It is a highly interdisci-
plinary field in which researchers from biology, medicine, engineering, computer
science, physics, and mathematics, among others, work together to provide the
best possible image. Imaging science is profoundly mathematical and challenging
from the modeling and the scientific computing point of view [1–5].
There are at least two major areas in imaging science in which applied math-
ematics has a strong impact: image processing, and image reconstruction. In
image processing the input is a (digital) image such as a photograph, while in
image reconstruction the input is a set of data. In the latter case, the data is
limited, and its poor information content is not enough to generate an image to
start with.
Image processing and analysis of information in images are methods that
become increasingly important in many technical and scientific fields. Image
processing techniques treat an image and apply numerical algorithms to either
improve the given image or to extract different features of it. Image segmentation
2 M. Moscoso
is typically used for the latter purpose. It refers to the process of partitioning an
image into multiple regions (locating objects and boundaries) in order to simplify
its representation for its further analysis. Each region shares the same properties
or characteristics such as color, intensity or texture. Many different techniques
have been applied for image segmentation. We mention here, graph partitioning
methods in which the image is modeled as a graph, level-sets methods in which
an initial shape is evolved towards the object boundary, or statistical methods
in which we view a region of an image as one realization of the various random
processes involved in the formation of that region (probability distribution func-
tions and histograms are used to estimate the characteristics of the regions). We
will not discuss the mathematics of image processing here.
Image reconstruction refers to the techniques used to create an image of the
interior of a body (or region) non-invasively, from data collected on its bound-
ary [1–5]. Image reconstruction can be seen as the solution of a mathematical
inverse problem in which the cause is inferred from the effect. As a consequence,
measurement and recording techniques designed to produce the images depend
deeply on the application we consider. In medicine, for example, very differ-
ent procedures are applied depending on the disease the physician is looking
for. Among these imaging techniques there are now in routine use, computer-
ized tomography, magnetic resonance, ultrasound, positron emission tomogra-
phy (PET), or electroencephalography (EEG), among others. They all acquire
data susceptible to be represented as maps containing positional information of
medical interest but, while each procedure operates on a different physical back-
ground principle and is sensitive to a different characteristic of the human body
(X-ray attenuation, dielectric properties, hydrogen nucleus density, reflection by
tissue interfaces, . . . ), they all share the same mathematical formulation of the
problem and similar numerical algorithms that allow the reconstruction of the
desired image.
We want to stress here, though, that it is always necessary a deep under-
standing of a mathematical model with which to interpret the measurements
and design numerical algorithms to reconstruct the desired images. Progress can
hardly be carried out without efficient and well designed numerical solvers. Imag-
ing is more than showing that an inverse problem may have a unique solution
under circumstances that are rarely satisfied. Modern imaging approaches deal
with understanding the trade off between data size, the quality of the image, the
computational complexity of the forward model used to generate the measure-
ments, and the complexity and stability of the numerical algorithm employed to
obtain the images. One neither has all the data he wants, nor can solve a very
general forward model to invert the data.
The purpose of this contribution is two fold: to give a first insight into dif-
ferent applications in which imaging has led to an important development, and
to outline the close mathematical connections between the imaging techniques
used in these applications. In Sect. 2 we review the impact that imaging has had
in medicine, geophysics and industry. In Sect. 3 we discuss the main ideas about
stability and resolution in imaging to introduce the reader to the basic notions
Introduction to Image Reconstruction 3
2 Applications
The problem of reconstructing images from boundary data is of fundamental
importance in many areas of science, engineering and industry. We will briefly
describe applications in different areas in which imaging has seen a significant
growth and innovation over the last decades. The presentation below has been
organized in three subsections: medicine, geophysics, and industry applications.
This organization is, however, somehow artificial. Seismic imaging, for example,
is described within the Geophysics subsection, but it is clear that seismic imaging
is of fundamental importance in the oil industry. Many of the imaging techniques
presented in each subsection have been used along the years, in one form or
another, within different applications as well. Magnetic resonance imaging, for
example, is described in the Medicine subsection since it is primarily used to
produce high quality images of the inside of the human body. However, magnetic
resonance is also used (in a different setting), for example, in oil exploration and
production since it can provide good estimates of permeability and fluid flow.
2.1 Medicine
X-ray Tomography
One of the main revolutions which medicine has experienced during the last few
decades began in 1972 with the first clinical application of X-ray tomography or
computerized tomography [2–5]. However, the theoretical foundations underly-
ing these image reconstruction technique are due to the mathematician Johan
Radon, back in 1917.
4 M. Moscoso
Object
Source
Film
Fig. 1. An X-ray source illuminates a 2D object from a single view. Measurements,
as line integrals of the attenuation coefficient over each line, are registered on the film
plane. Observe that depth information is lost
first applications of Radon’s theory were done by Bracewell in 1956 in the con-
text of radioastronomy, but his work had little impact and was not known in the
medical world. Both, Cormack and Hounsfield were awarded the Nobel Prize for
medicine in 1979. The new and powerful computers needed for the implementa-
tion of the algorithms, were crucial for the success of this technology.
A description of the theory of imaging in X-ray tomography can be found in
the lecture by Frank Natterer in this book. He surveys integral geometry and re-
lated imaging techniques. Computerized tomography reconstruction algorithms,
such as the filtered backprojection algorithm, Kaczmarz’s method, or Fourier
methods, are also presented.
While X-ray tomography uses ionizing radiation to form images, magnetic res-
onance imaging (MRI) uses non-ionizing signals. Magnetic resonance was orig-
inally named nuclear magnetic resonance, but the word nuclear was dropped
because of the negative connotations associated with the word nuclear. Neither
people nor local politicians wanted anything related to nuclear in their com-
munities. The term nuclear simply referred to the fact that all atoms have a
nucleus. The measurements are taken under powerful magnetic fields in con-
junction with radio waves. When placed in an external magnetic field, particles
with spin different than zero, such as protons, can absorb energy that is later
reemitted producing secondary radio waves. The detected signal results from the
difference between the absorbed and the emitted energies.
Magnetic resonance imaging, proposed by Richard Ernst in 1975, is proba-
bly the major competitor to X-ray tomography. In 1980 Edelstein and coworkers
demonstrated imaging of the body with his technique. Richard Ernst was re-
warded for his contributions to the development of high resolution techniques
in magnetic resonance imaging with the Nobel Prize in Chemistry. At the be-
ginning of the 90s magnetic resonance imaging started to be used for functional
imaging purposes. This new application allows the study of the function of the
various regions of the human brain. In 2003, Paul C. Lauterbur and Sir Peter
Mansfield were awarded the Nobel Prize in Medicine for their seminal discoveries
concerning the use of magnetic resonance to visualize different structures which
led to revolutionary insights into the functions of the brain and the workings of
the human body.
Besides its ability for functional imaging, magnetic resonance imaging is also
better suited for soft tissue than X-ray tomography. It can image different tis-
sue properties by variation of the scanning parameters, and tissue contrast can
be changed and enhanced in various ways to detect different features. The dis-
advantages of magnetic resonance imaging over X-ray tomography is that it
is more expensive and more time consuming. Nevertheless, the imaging tech-
niques behind this modality are the same used to reconstruct images in X-ray
tomography.
6 M. Moscoso
Can we use light to see or image inside the body? Diffuse optical tomography in
the near-infrared is an emerging modality with potential in medical diagnosis and
biomedicine. A near-infrared light source impinges upon the surface and the light
propagates through the body (Fig. 2). The light that emerges is detected at an
array and used to infer the local optical properties (absorption and scattering) of
the illuminated tissue. The major absorbers when using near-infrared light are
water and both oxygenated and deoxygenated hemoglobin. Cancer metastasis
requires the growth of a new network of blood vessels (angiogenesis). The greater
supply of blood around the tumors give rise to an absorbing obstacle feasible to
be detected by this technique.
Another important current application of diffuse optical tomography is brain
imaging. By shining near-infrared light on the scalp, changes in neural activ-
ity can be measured because of the tight correlation existing between brain
activation and the changes in concentration of oxygenated and deoxygenated
hemoglobin (cerebral hemodynamics). This is possible due to the relationship of
the absorption spectra of oxygenated hemoglobin and deoxygenated hemoglobin
at near-infrared wavelengths. In addition, diffuse optical tomography can detect
and localize important events such as hemorrhagic strokes in the brain.
We would like to emphasize that to reconstruct quantitative images of the
absorption and scattering parameters from near-infrared light measurements one
Object
Near−infrared
source
Detectors
Fig. 2. A near-infrared source illuminates a 2D object. The photons do not travel
along straight lines. This causes blurring
Introduction to Image Reconstruction 7
1 ∂I(x, k̂, t)
+ k̂ · ∇x I(x, k̂, t) + (Σ (s) (x) + Σ (a) (x))I(x, k̂, t) =
v ∂t
Σ (s) (x) P (k̂ · k̂ )I(x, k̂ , t) dk̂ , (1)
where v, Σ (s) , and Σ (a) are the velocity of the light in the medium, the scat-
tering coefficient, and the absorption coefficient, respectively. The fundamental
quantity of radiative transfer is the specific intensity I which depends on the
position vector x and unit direction vector k̂. The phase function P (k̂ · k̂ ) in (1)
describes the directional distribution of light that scatters in direction k̂ due to
light of unit energy density in direction k̂ . We have considered here the scalar
radiative transport equation.
The radiative transport equation (1) takes into account absorption and scat-
tering due to the inhomogeneities in tissue. Analytical solutions to this inte-
grodifferential equation are known only for relatively simple problems. For more
complicated problems numerical solutions are needed. The use of this theory
for imaging purposes in biological tissue can be found in the lecture by Miguel
Moscoso. A Monte Carlo method to solve this equation can also be found.
Before we proceed, we mention that one can use the diffusion equation to
model light propagation in tissue as an alternative to the more complicated
radiative transport equation. The diffusion equation is much easier to solve but
has limitations such that it does not describe accurately light propagation in
optically thin media nor light near sources and boundaries. Nevertheless, the
diffusion approximation to the radiative transport equation is widely used in the
literature.
2.2 Geophysics
Seismic Imaging
Seismic surveying involves sending sound waves underground and recording the
reflected signals from the earth’s interior [1]. Reflection occurs when the waves
hit a change in medium and experience a change in their velocity. This is a
simple fact. But changes in the earth’s properties take place in very different
spatial scales and are not easy to model. As a first approximation the earth is a
stratified medium in which the thickness of the layer might vary from less than
a meter to several hundreds of meters. On the other hand, tectonic forces in the
earth can bend these layers or can give rise to fractures. Besides, the interior
structure can be severely affected by the influence of other processes like the
extreme heat generated from the earth’s interior at some locations. As a result,
the seismic waves propagating through the earth will be refracted, reflected,
and diffracted from the heterogeneous medium, and the recorded signal will
show a very complicated structure. Earth scientists use these data to create a
Introduction to Image Reconstruction 9
Electromagnetic Imaging
but have a larger penetration range. Higher frequencies provide better resolu-
tion but limit the range of the measurements. The chapter by Oliver Dorn, Hugo
Bertete-Aguirre, and George Papanicolaou presents this modality of imaging and
a method for solving the inverse Maxwell problem in three dimensions.
More recent method involve deployment of both vertical borehole arrays and
large surface arrays.
2.3 Industry
Nondestructive Testing
The problem in which one needs to detect the presence of fissures or flaws from
measurements on the surface of a material is of much interest in industry. Nonde-
structive testing is any examination procedure performed on an object (without
destroying it) in order to detect the presence of defects such as cracks that can
alter its usefulness [3]. Nondestructive testing has become, in fact, an integral
part of almost all industrial processes in which a product failure might result in
an accident or body injury. For example, in aircraft maintenance it is important
to inspect the mechanical damage and to evaluate the amount of work necessary
to repair it. It is also important to do it in an efficient way, as the maintenance of
aircraft must be accomplished within a scheduled time. Nondestructive testing
can accomplish the inspection and can detect cracks or any other irregularities
in the airframe structure that are not visible. In other words, nondestructive
testing is the technology which provides information regarding the condition
of the material under examination once the necessary measurements have been
performed.
It is difficult to say when or who performed the first nondestructive testing.
Nondestructive tests can actually be performed in a very natural way without
employing sophisticated technologies. Listening to the sound of different metals
has been used along the centuries to determine the quality of the piece being
shaped (a sword, a ring, a bell, . . . ). However, it was in the late 1950s when
nondestructive techniques started to experience an important growth through
the development of new technologies and ideas. The mathematical study of this
inverse problem was initiated by A. Friedman and M. Vogelius in 1989.
There is a broad range of techniques that are applied nowadays depending
on the kind of measured data and the specific industrial situation under consid-
eration. Boundary data can be obtained from thermal, acoustic, elastostatic, or
electrical measurements. Lately, hybrid methods that perform different kind of
Introduction to Image Reconstruction 11
History Matching
Reservoir simulators are routinely used to predict and optimize oil production.
The accuracy of the numerical solution computed by the simulation depends
on the accuracy of the physical parameters that characterize the reservoir and,
unfortunately, these are poorly known due to the relative inaccessibility of the
reservoir to sampling. Therefore, proper characterization of the petroleum reser-
voir heterogeneity is a crucial aspect of any optimal reservoir management strat-
egy. It helps to better understand the reservoir behavior so that its performance
can be predicted and controlled with higher reliability. The history matching
problem consists in adjusting a set of parameters, such as the permeability and
porosity distributions, in order to match the data obtained with the simulator
to the actual production data in the reservoir. The input data used to solve the
inverse problem in history matching usually consist of the pressure or the flow
at the production wells.
One important difference between the history matching problem and other
applications such as X-ray tomography or magnetic resonance imaging is that in
the history matching problem the experimental setup cannot easily be changed
in order to obtain independent data. Typically, only one field experiment is
available due to the production process. A small reservoir model might have
50,000 to 100,000 grid blocks and solutions to the history matching problem
are far from being unique. As a consequence, production data rarely suffices to
characterize reservoir heterogeneity and it is, therefore, desirable to integrate all
12 M. Moscoso
other available data into the model, such as geological or seismic interpretations,
to reduce the amount of uncertainty. Lately, several researchers are making a lot
of effort to incorporate geological data, coming from sparse measurements of the
permeability, as constrains to the inverse problem. Most of the approaches are
formulated to honor the histograms generated from these data.
3 Stability
Let us consider the Hilbert spaces P and D (the parameter and data spaces,
respectively) and the linear bounded operator A : P → D. The goal of any
imaging technique is to take a set of data d ∈ D and produce an image of some
(unknown) physical quantity of interest f ∈ P . Mathematically, this is done by
solving an inverse problem represented in abstract form by
Af = d. (2)
However, the inverse problems arising from image reconstruction are usually
ill-posed in the sense of Hadamard. A problem is well-posed in the sense of
Hadamard if:
1. A solution exists.
2. The solution is unique.
3. The solution depends continuously on the given data.
Otherwise, the problem is ill-posed. A good measure of the degree of ill-posedness
is given by the singular value decomposition (SVD) of A. The faster the decay
rate of the singular values, the worse the degree of ill-posedness. If the singu-
lar values decay asymptotically at an exponential rate, the inverse problem is
severely ill-posed. This is the case, for example, of diffuse optical tomography or
cross-borehole electromagnetic induction tomography.
The most important issue in image reconstruction, from the computational
point of view, is the failure of the third property. One can deal with the first two
by approximate inverses such as the well known Moore–Penrose pseudoinverse
A† , but if the third property is not fulfilled any simple image reconstruction
algorithm will lead to instabilities and the image will not be reliable. To impose
stability on an ill-posed inverse problem we must introduce continuity in the
inverse operator A−1 by means of a regularization method [4].
We stress here that any measurement procedure leads to data perturbed by
noise. Rather than attempting to invert A directly (perfect inverses in realis-
tic problems, especially those arising in imaging reconstruction, typically never
happen), we adopt a least square approach and seek a solution f that minimizes
the cost functional
1
J(f ) = ||R(f )||22 , (3)
2
where R(f ) = A f − d is the residual operator which measures the data misfit.
Introduction to Image Reconstruction 13
over s ∈ R is available for some projections θ ∈ [0, π], where δ is the Dirac’s delta
function. The inverse problem is to find f from a finite number of measurements
(Rf )(θi , sj ) = dij . If Rf is available for all θ and s (without noise), then the
inverse Radon transform solves the problem exactly. This inverse problem is,
nevertheless, ill posed.
Simple backprojection is conceptually quite easy but it does not correctly
solve this inverse problem. Simple backprojection reconstructs an image by tak-
ing each projection and smearing (or backprojecting) it along the line it was
Introduction to Image Reconstruction 15
Acknowledgements
References
1. Claerbout J.F., Fundamentals of Geophysical Data Processing, McGraw-Hill
international series in the earth and planetary sciences, 1976.
2. Epstein C.L., Introduction to the Mathematics of Medical Imaging, Prentice Hall,
2003.
3. Hellier C., Handbook of nondestructive evaluation, McGraw-Hill, 2001.
4. Kirsch A., An introduction to the mathematical theory of inverse problems, Applied
Mathematical Science vol. 120, Springer, 1996.
5. Natterer F. and Wübbeling F., Mathematical Methods in Image Reconstruction,
SIAM Monographs on Mathematical Modeling and Computation, 2001.
X-ray Tomography
Frank Natterer
1 Introduction
Since the advent of CT many imaging techniques have come into being, some
of them being only remotely related to the straight line paradigm of CT, such
as impedance tomography, optical tomography, and ultrasound transmission to-
mography. For the study of these advanced technique a thorough understanding
of CT is at least useful. Many of the fundamental tools and issues of CT, such as
backprojection, sampling, and high frequency analysis, have their counterparts
in the more advanced techniques and are most easily in the framework of CT.
The outline of the paper is as follows. We start with a short account of the
relevant parts of integral geometry, with the Radon transform as the central tool.
Then we discuss in some detail reconstruction algorithms, in particular the nec-
essary discretizations for methods based on inversion formulas and convergence
properties of iterative algorithm. Finally we concentrate on the 3D case which
is the subject of current research.
2 Integral Geometry
In this section we introduce the relevant integral transforms, derive inversion
formulas, and study the ranges. For a thorough treatment see Gelfand, Graev,
and Vilenkin (1965), Helgason (1999), Natterer (1986).
Theorem 2.1.
(Rf )∧ (θ, σ) = (2π)(n−1)/2 fˆ(σθ) .
Note that the Fourier transform on the left hand side is the 1D Fourier
transform of Rf with respect to its second variable, while the Fourier transform
on the right hand side is the nD Fourier transform of f .
For the proof of Theorem 2.1 we make use of the definition (2) of Radon
transform, yielding
(Rf )∧ (θ, σ) = (2π)−1/2 e−isσ (Rf )(θ, s)ds
R1
= (2π)−1/2 e−isσ f (sθ + y)dyds .
R1 θ⊥
= (2π)(n−1)/2 fˆ(σθ) .
The proofs of many of the following results follow a similar pattern. So we omit
proofs unless more sophisticated tools are needed.
It is possible to extend R to a bounded operator R : L2 (Rn ) → L2 (S n−1 ×R1 ).
As such R has an adjoint R∗ : L2 (S n−1 × R1 ) → L2 (Rn ) which is easily seen
to be
(R∗ g)(x) = g(θ, x · θ)dθ . (4)
S n−1
∗
R is called the backprojection operator in the imaging literature.
We also will make use of the Hilbert transform
1 f (t)
(Hf )(s) = dt (5)
π s−t
which, in Fourier domain, is given by
(Hf )∧ (σ) = −i sgn(σ)fˆ(σ) (6)
with sgn(σ) the sign of the real number σ. Now we are ready to state and to
prove Radon’s inversion formula:
For the proof we write the Fourier inversion formula in polar coordinates and
make use of Theorem 2.1 and the evenness of g, i.e. g(θ, s) = g(−θ, −s).
Special cases of Theorem 2.2 are
1 g (θ, s)
f (x) = dsdθ (7)
4π 2 x·θ−s
S 1 R1
for n = 2 and
1
f (x) = − g (θ, x · θ)dθ (8)
8π 2
S2
for n = 3. Note that there is a distinctive difference between (7) and (8): The
latter one is local in the sense that for the reconstruction of f at x0 only the
integrals g(θ, s) over those planes x · θ = s are needed that pass through x0 or
nearby. In contrast, (7) is not local in this sense, due to the integral over R1 .
In order to study the stability of the inversion process we introduce the
Sobolev spaces H α on Rn , S n−1 × R1 with norms
f 2H α (Rn ) = (1 + |ξ|2 )α |fˆ(ξ)|2 dξ ,
Rn
gH α (Sn−1 ×R1 ) =
2
(1 + σ 2 )α |ĝ(θ, σ)|2 dσdθ .
S n−1 R1
Theorem 2.3. There exist constants c, C > 0 depending only on α, n, such that
cf H α (Rn ) ≤ Rf (S n−1 ×R1 ) ≤ Cf H α (Rn )
H α+(n−1)/2
P is called the ray (or X-ray) transform. The projection-slice theorem for P reads
Theorem 2.5.
(P f )∧ (θ, ξ) = (2π)1/2 fˆ(ξ), ξ ∈ θ⊥ .
The Fourier transform on the left hand side is the (n − 1)D Fourier transform
in θ⊥ , while the Fourier transform on the right hand side is the nD Fourier
transform in Rn .
The adjoint ray transform, which we call backprojection again, is given by
(P ∗ g)(x) = g(θ, Eθ x)dθ (10)
S n−1
Theorem 2.7. Let n = 3, and assume that each great circle on S 2 meets S0 .
Then, f is uniquely determined by g(θ, ·), θ ∈ S0 .
The condition on S0 in Theorem 2.7 is called Orlov’s completeness condition.
In the light of Theorem 2.5, the proof of Theorem 2.7 is almost trivial: Let
ξ ∈ R3 \ {0}. According to Orlov’s condition there exists θ ∈ S0 on the great
circle ξ ⊥ ∩ S 2 . For this θ, ξ ∈ θ⊥ , hence
fˆ(ξ) = (2π)−1/2 ĝ(θ, ξ)
by Theorem 2.5. Thus fˆ is uniquely (and stably) determined by g on S0 .
22 F. Natterer
The relevant integral transform for fully 3D X-ray tomography is the cone beam
transform in R3
∞
(Cf )(a, θ) = f (a + tθ)dt (11)
0
where θ ∈ S . The main difference to the ray transform P is that the source
2
Theorem 2.8.
∂ ∂
(Rf )(θ, s) |s=a·θ = (Cf )(a, ω)dω .
∂s ∂θ
θ ⊥ ∩S 2
The notation on the right hand side needs explication: (Cf )(a, ·) is a function
on S 2 . θ is a tangent vector to S 2 for each ω ∈ S 2 ∩ θ⊥ . Thus the derivative ∂θ
∂
makes sense on S 2 ∩ θ⊥ .
Since Theorem 2.8 is the basis for present day’s work on 3D reconstruction
we sketch two proofs. The first one is completely elementary. Obviously it suffices
to put θ = e3 , the third unit vector, in which case it reads
⎡ ⎛ ⎛ ⎞⎞⎤
∂ ∂ ω
(Rf )(θ, a3 ) = ⎣ (Cf ) ⎝a, ⎝ ⎠⎠⎦ dω
∂s ∂z z
ω∈S 1 z=0
where a3 is the third component of the source point a. It is easy to verify that
right- and left-hand side both coincide with
⎛ ⎛ ⎞⎞
∂f ⎝ x
a + ⎝ ⎠⎠ dx .
∂x3 0
R2
Theorem 2.9. Let the source curve A satisfy the following condition: Each plane
meeting supp (f ) intersects A transversally. Then, f is uniquely (and stably)
determined by (Cf )(a, θ), a ∈ A, θ ∈ S 2 .
∂ ∂
Rf (θ, s) = Rf (θ, s)
∂s s=x·θ ∂s s=a·θ
∂
= (Cf )(a, ω)dω
∂θ
θ ⊥ ∩S 2
This is Novikov’s inversion formula; see Novikov (2000). Note that the back-
∗
projection R−µ is applied to a vector valued function. For µ = 0 Novikov’s
formula reduces to Radon’s inversion formula (7).
There is also an extension of Theorem 2.4 to the attenuated case:
Theorem 2.11. Let k > m ≥ 0 integer and h the function from Theorem 2.10.
Then,
sm eikϕ+h(θ,s) (Rµ f )(θ, s)dθds = 0
R1 S 1
⎛ ⎞
cos ϕ
where θ = ⎝ ⎠.
sin ϕ
The proofs of Theorems 2.10 and 2.11 are quite involved. They are both
based on the following remark: Let
u(x, θ) = u (x)eiϕ
>0 odd
⎛ ⎞
cos ϕ
with certain functions u (x) and θ = ⎝ ⎠.
sin ϕ
Of course this amounts to saying that u(x, ·) admits an analytic continuation
from S 1 into the unit disk.
X-ray Tomography 25
Now let f be a vector field. The scalar transforms P , R can be applied compo-
nentwise, so P f , Rf make sense. They give rise to the vectorial ray transform
Obviously, vectorial transforms can’t be invertible. In fact their null spaces are
huge.
Hence θ⊥ is invariant under fˆ. This is only possible if fˆ(ξ) = iψ̂(ξ)ξ with some
ψ̂(ξ). It remains to show that ψ̂ is sufficiently regular to conclude that f = ∇ψ.
A similar proof leads to
3 Reconstruction Algorithms
There are essentially three classes of reconstruction algorithms. The first class
is based on exact or approximate inversion formulas, such as Radon’s inversion
formula. The prime example is the filtered backprojection algorithm. It’s not only
the work horse in clinical radiology, but also the model for many algorithms in
other fields. The second class consists of iterative methods, with Kaczmarz’s
method (called algebraic reconstruction technique (ART) in tomography) as
prime example. The third class is usually called Fourier methods, even though
Fourier techniques are used also in the first class. Fourier methods are direct
implementations of the projection slice theorem (Theorems 2.1, 2.5) without any
reference to integral geometry.
The proof of Theorem 3.1 require not much more than the definition of R∗ .
Theorem 3.1 is turned into an (approximate) inversion formula for R by
choosing V ∼ δ. Obviously V is the result of the reconstruction procedure for
f = δ. Hence V is the point spread function, i.e. the function the algorithm
would reconstruct if the true f were the δ-function.
1
v̂(σ) ∼ (2π)−n+1/2 σ n−1 .
2
Let φ be a low pass filter, i.e. φ(σ) = 0 for |σ| > 1 and, for convenience, φ
even. Let Ω be the (spatial) bandwidth. By Shannon’s sampling theorem Ω
corresponds to the (spatial) resolution 2π/Ω:
X-ray Tomography 27
Theorem 3.3. Let f be Ω-bandlimited, i.e. fˆ(ξ) = 0 for |ξ| > Ω. Then, f is
uniquely determined by f (h), ∈ Zn , h ≤ π/Ω. Moreover, if h ≤ 2π/Ω, then
f (x)dx = hn f (h) .
Rn
Then,
Ω2 1 s 2
v(s) = u(Ωs) , u(s) = sinc(s) − sinc .
4π 2 2 2
Now we describe the filtered backprojection algorithm for the reconstruction of
the function f from g = Rf . We assume that f has the following properties:
where s = ∆s and ∆s ≤ π/Ω. Equation (15) is the filtering step in the filtered
backprojection algorithm. In the backprojection step we compute R∗ (v ∗ g) in a
discrete setting. We restrict ourselves to the case n = 2, i.e.
28 F. Natterer
(R∗ (v ∗ g))(x) = (v ∗ g)(θ, x·)dθ .
S1
In order to discretize this integral properly we make use of Theorem 3.3 again.
For this we⎛have to⎞determine the bandwidth of the function ϕ → (v ∗ g)(θ, x · θ)
cos ϕ
where θ = ⎝ ⎠.
sin ϕ
2π Ω
−ikϕ ik −ikψ
e (v ∗ g)(θ, x · θ)dϕ = f (y)e v̂k (σ)Jk (−σ|x − y|)dσdy
2π
0 |y|<ρ −Ω
where ψ = arg(y) and Jk is the Bessel function of order k of the first kind.
2π
i−k
Jk (s) = e−ikϕ+s cos ϕ dϕ
2π
0
Rj f = gj , j = 1, . . . , p (17)
X-ray Tomography 29
where Rj are certain linear operators from a Hilbert space into an Hilbert space
Hj . For instance, for the 2D problem we have
Kaczmarz’s method solves linear systems of the kind (18) by successively pro-
jecting orthogonally onto the affine subspaces Rj f = gj of H:
fj = Pj modp fj−1 , j = 1, 2, . . . .
Theorem 3.5. Let (17) be consistent. Assume that Cj ≤ Rj Rj∗ (i.e. Cj her-
mitian and Cj − Rj Rj∗ positive semidefinite) and 0 < ω < 2. Then, fj =
Pjωmodp fj−1 converges for f0 = 0 to the solution of (18) with minimal norm.
The condition 0 < ω < 2 is reminiscent of the SOR (successive overre-
laxation) method of numerical analysis. In fact Kaczmarz’s method can be
viewed as an SOR method applied to the linear system R∗ Rf = R∗ g where
R = R 1 × · · · × Rp .
Theorem 3.5 can be applied directly to tomographic problems. The conver-
gence behaviour depends critically on the choice of ω and of the ordering of (17).
We analyse this for the standard case (18). For convenience we assume that f is
supported in |x| < 1 and gj = 0, j = 1, . . . , p.
The following result is due to Hamaker and Solmon (1978):
Pj Cm ⊆ Cm .
P ω = P1ω · · · Ppω .
ω = 0.5
ω = 1.0
0 m
0 30
ρm (ω)
1 ω = 0.1
ω = 0.5
ω = 1.0
0 m
0 30
ρm (ω)
1 ω = 0.1
ω = 0.5
ω = 1.0
0 m
0 30
First we consider the solid lines. They correspond to ω = 1. For the sequential
ordering ϕj = jπ/p, j = 1, . . . , p the values of ρm (ω) are displayed in the top
figure. We see that ρm (ω) is fairly large for low order subspaces Cm , indicating
slow convergence for low frequency parts of f , i.e. for the overall features of f .
The other figures shows the corresponding values of ρm (ω) for a non-consecutive
ordering of the ϕj suggested by Herman and Meyer (1993) and for a random
ordering respectively. Both orderings yield much smaller values of ρm (ω), in
X-ray Tomography 31
particular on the low order subspaces. The conclusion is that for practical work
one should never use the consecutive ordering, and random ordering is usually
as good as any other more sophisticated ordering.
In particular in emission tomography the use of multiplicative iterative meth-
ods is quite common. The most popular of these multiplicative algorithms is the
expectation-maximization (EM) algorithm: Let Af = g be a linear system, with
all the elements of the (not necessarily square) matrix A being non-negative.
Assume that A is normalized such that A1 = 1 where 1 stands for vectors of
suitable dimension containing only 1 s. The EM algorithm for the solution of
Af = g reads
g
f j+1 = f j A∗ , j = 1, 2, . . . ,
Af j
where division and multiplication are understood componentwise. One can show
that f j converges to the minimizer of the log likelihood function
There exists also a version of the EM algorithm that decomposes the system
Af = g into smaller ones, Aj f = gj , j = 1, . . . , p, as in the Kaczmarz method
(ordered subset EM, OSEM; see Hudson and Larkin (1992)). As in the Kaczmarz
method one can obtain good convergence properties by a good choice of the
ordering, and the convergence analysis is very much the same.
For more on iterative methods in CT see Censor (1981).
Fourier methods make direct use of relations such as Theorem 2.1. Let’s consider
the 2D case in which we have
One of the most urgent needs of present day’s X-ray CT is the development
of accurate and efficient algorithms for the reconstruction of f from (Cf )(a, θ)
where a is on a source curve A (typically a helix) and θ ∈ S 2 . In practice, θ is
restricted to a subset of S0 , but we ignore this additional difficulty.
We assume that the source curve A satisfies Tuy’s condition, i.e. A intersects
each plane hitting supp (f ) transversally. Let A be the curve x = a(λ), λ ∈ Λ.
Tuy’s condition implies that for each s with (Rf )(θ, s) = 0 there is λ such that
s = a(λ) · θ. In fact there may be more than one such λ, but we assume that
their number is finite. Then we may choose a function M (θ, s) such that
M (θ, λ) = 1
λ
s=a(λ)·θ
for each s. Let g(λ, θ) = (Cf )(a(λ), θ) be the data function. Put
∂
G(λ, θ) = g(λ, ω)dω .
∂θ
θ ⊥ ∩S 2
Then we have
For the proof we start out from the 3D case of Theorem 3.1:
(V ∗ f )(x) = w (x · θ − s)(Rf )(θ, s)dsdθ
S 2 R1
= w(x · θ − s)(Rf ) (θ, s)dsdθ .
S 2 R1
the factor M (θ, s) being due to the fact that λ → a(λ) · θ is not one-to-one. By
Theorem 2.8,
(Rf ) (θ, a(λ) · θ) = G(λ, θ) ,
and this finishes the proof.
X-ray Tomography 33
Theorem 4.1 is the starting point for a filtered backprojection algorithm. For
1
w = − δ ,
8π 2
δ the 1D δ-function, we have V = δ, the 3D δ-function, hence
x − a(λ)
f (x) = |x − a(λ)|−2 Gw λ, dλ ,
|x − a(λ)|
Λ
1
G (λ, ω) = − 2
w
δ (ω · θ)G(λ, θ)|a (λ) · θ|M (θ, λ)dθ, ω ∈ S 2 .
8π
S2
This is the formula of Clack and Defrise (1994) and Kudo and Saito (1994). For
more recent work see Katsevich (2002).
There exist various approximate solutions to the cone-beam reconstruction
problem, most prominently the FDK formula for a circular source curve (which
doesn’t suffice Tuy’s condition). It can be viewed as a clever adaption of the 2D
filtered backprojection algorithm to 3D; see Feldkamp, Davis and Kress (1984).
An approximate algorithm based on Orlov’s condition of Theorem 2.7 is the π
method of Danielsson et al. (1999).
References
1. M. Abramowitz and I.A. Stegun (1970): Handbook of Mathematical Functions.
Dover.
2. G. Beylkin (1995): ‘On the fast Fourier transform of functions with singularities’,
Appl. Comp. Harm. Anal. 2, 363-381.
3. Y. Censor (1981): ‘Row–action methods for huge and sparce systems and their
applications’, SIAM Review 23, 444-466.
4. R. Clack and M. Defrise (1994): ‘A Cone–Beam Reconstruction Algorithm Using
Shift–Variant Filtering and Cone–Beam Backprojection’, IEEE Trans. Med. Imag.
13, 186-195.
5. P.E. Danielsson, P. Edholm, J. Eriksson, M. Magnusson Seger, and H. Turbell
(1999): The original π-method for helical cone-beam CT, Proc. Int. Meeting on
Fully 3D-reconstruction, Egmond aan Zee, June 3-6.
6. L.A. Feldkamp, L.C. Davis, and J.W. Kress (1984): ‘Practical cone–beam algo-
rithm’, J. Opt. Soc. Amer. A 6, 612-619.
7. K. Fourmont (1999): ‘Schnelle Fourier–Transformation bei nicht–äquidistanten
Gittern und tomographische Anwendungen’, Dissertation Fachbereich Mathematik
und Informatik der Universität Münster, Münster, Germany.
8. I.M. Gel‘fand, M.I. Graev, N.Y. Vilenkin (1965): Generalized Functions, Vol. 5:
Integral Geometry and Representation Theory. Academic Press.
9. P. Grangeat (1991): ‘Mathematical framework of cone–beam reconstruction via
the first derivative of the Radon transform’ in: G.T. Herman, A.K. Louis and
F. Natterer (eds.): Lecture Notes in Mathematics 1497, 66-97.
10. C. Hamaker, K.T. Smith, D.C. Solmon and Wagner, S.L. (1980): ‘The divergent
beam X-ray transform’, Rocky Mountain J. Math. 10, 253-283.
34 F. Natterer
11. C. Hamaker and D.C. Solmon (1978): ‘The angles between the null spaces of
X-rays’, J. Math. Anal. Appl. 62, 1-23.
12. S. Helgason (1999): The Radon Transform. Second Edition. Birkhäuser, Boston.
13. G.T. Herman (1980): Image Reconstruction from Projection. The Fundamentals of
Computerized Tomography. Academic Press.
14. G.T. Herman and L. Meyer (1993): ‘Algebraic reconstruction techniques can be
made computationally efficient’, IEEE Trans. Med. Imag. 12, 600-609.
15. H.M. Hudson and R.S. Larkin (1994): ‘Accelerated EM reconstruction using or-
dered subsets of projection data’, IEEE Trans. Med. Imag. 13, 601-609.
16. A.C. Kak and M. Slaney (1987): Principles of Computerized Tomography Imaging.
IEEE Press, New York.
17. A. Katsevich (2002): ‘Theoretically exact filtered backprojection-type inversion
algorithm for spiral CT’, SIAM J. Appl. Math. 62, 2012-2026.
18. H. Kudo and T. Saito (1994): ‘Derivation and implementation of a cone–beam
reconstruction algorithm for nonplanar orbits’, IEEE Trans. Med. Imag. 13,
196-211.
19. F. Natterer (1986): The Mathematics of Computerized Tomography. John Wiley &
Sons and B.G. Teubner. Reprint SIAM 2001.
20. F. Natterer and F. Wübbeling (2001): Mathematical Methods in Image Reconstruc-
tion. SIAM, Philadelphia.
21. F. Natterer and E.L. Ritman (2002): ‘Past and Future Directions in X-Ray Com-
puted Tomography (CT)’, to appear in Int. J. Imaging Systems & Technology.
22. R.G. Novikov (2000): ‘An inversion formula for the attenuated X-ray transform’,
Preprint, Departement de Mathématique, Université de Nantes.
23. V.A. Sharafutdinov (1994): Integral geometry of Tensor Fields. VSP, Utrecht.
24. G. Steidl (1998): ‘A note on fast Fourier transforms for nonequispaced grids’, Ad-
vances in Computational Mathematics 9, 337-352.
25. S. Webb (1990): From the Watching of Shadows. Adam Hilger.
Adjoint Fields and Sensitivities
for 3D Electromagnetic Imaging in Isotropic
and Anisotropic Media
for example Davis and Annan [DA89] or Fisher et al. [FMA92]. Here the probing
waves are normally in the radio to microwave range of frequencies. GPR meth-
ods for imaging conductivity use comparatively high frequencies (≥ 100 MHz) in
order to have a signal that propagates through the Earth as a wave. The main
practical difficulty with GPR imaging is that at these frequencies its depth of
penetration is usually no deeper than about 5 m into the Earth in dry soil/sand,
while maximum depths of less than 1–2 m are more typical in wet earth due to
high local wave attenuation.
ElectroMagnetic Induction Tomography (EMIT) has been investigated re-
cently as a promising new tool for imaging electrical conductivity variations in
the earth (Wilt et al. [WAMBL, WMBTL] and Buettner et al. [BB99]) with
a depth of investigation greater than Ground Penetrating Radar (GPR). The
source is a magnetic field generated by currents in wire coils. This source field
is normally produced in one borehole, while the received signals are the mea-
sured small changes in magnetic field in another, more distant borehole (Fig. 1).
The method may also be used successfully in combination with surface sources
and receivers. The goal of this procedure is to image electrical conductivity
variations in the earth, much as X-ray tomography is used to image density
variations through cross-sections of the body. However, the fields tend to dif-
fuse rather than propagate as waves in this frequency band (from about 1 to
20 kHz) and the inversion techniques are therefore very different from those for
GPR. Although field techniques have been developed and applied previously to
collection of such EM data, the algorithms for inverting the magnetic field data
to produce the desired images of electrical conductivity have not kept pace. The
current state of the art in electromagnetic data inversion (Alumbaugh and Mor-
rison [AM95a, AM95b]) is based on the Born/Rytov approximation (requiring
source air
receivers
soil
Fig. 1. The figure shows a possible experimental setup for 3D-EMIT. Three boreholes
surround the area of interest. A z-directed dipole transmitter q is deployed in one of the
boreholes, and data are gathered in the other boreholes and at the surface at positions
dn , n = 1, . . . , N . From these data, we want to recover the (isotropic or anisotropic)
conductivity and permittivity distribution in the earth
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 37
wavelength of the radiating sources and the size of the aperture. When we are
considering electromagnetic fields in the diffusive regime, it becomes problem-
atic to define a useful resolution criterion. At very low frequencies such as those
used in Electrical Resistance/Impedance Tomography (ERT/EIT) there is no
wavelength associated with the signals and the resolution must instead be de-
fined in terms of the numbers of source/receiver electrode pairs and the size of
the region being imaged (see Borcea [Bor02] or Ramirez et al. [RDLOC]). The
mathematically analogous problem in hydrology has also been treated (see Vasco
et al. [VDL97]). For EMIT, the situation is closer to that of ERT/EIT than it is
to that of wave propagation problem like seismic, ground penetrating radar, or
optics. The wavelength does not provide a useful measure of the method’s true
resolution. A better approach in this situation is to define sensitivity functions
(see for example Spies and Habashy [SH95]). A short introduction into sensitivity
analysis for inverse problems is given in the following.
the propagating fields are now vector fields rather than scalar fields. This gives
us a richer variety of sensitivity functions, depending on the component of the
field which is used for collecting the data. We will see that the shapes of the
sensitivity functions in EMIT show more variations than in DOT. For certain
source-receiver geometries, the sensitivities in the region of interest are also here
mainly concentrated on some almost ‘banana-shaped’ regions, whereas for other
source-receiver constellations these regions look more like ‘elongated doughnuts’,
with almost vanishing sensitivities close to the connecting line of source and re-
ceiver. We mention that another related geophysical application, which EMIT
and DOT are interesting to compare with, is hydrology. That situation is treated
in more details for example in Vasco et al. [VDL97].
The nonlinear Kaczmarz-type approach for solving the inversion problem
(which can be characterized as a ‘single-step adjoint-field inversion method’)
will be presented further below in this paper. It can be viewed as a nonlinear
generalization of the algebraic reconstruction technique (ART) in X-ray tomogra-
phy [Nat86,NW01]. In contrary to the filtered backprojection method mentioned
earlier, ART uses only part of the data in each step of the reconstruction. It is
an iterative technique and applies successive corrections to some initial guess
for the parameter distribution, which in X-ray tomography is the attenuation.
Each of these corrections is calculated by ‘backprojecting’ a (preprocessed or
filtered) part of the residuals along those straight lines into the imaging domain
which correspond to the incoming direction of the probing X-rays. We will show
that, in an analogous way, each individual backprojection step of our generalized
ART algorithm for EMIT takes place along sensitivity functions. An important
difference of our generalized ART scheme to the classical ART algorithm is its
nonlinearity. Whereas the sensitivities in the classical ART method are always
concentrated on the same straight lines, the shapes of our sensitivity functions
depend on the latest best guess for the parameters and therefore change with
each update. The final reconstruction of our inversion method is given as a super-
position of sensitivity functions with weights determined by the residual values
which appear during the reconstruction process. The knowledge of the general
structure of the sensitivity functions during the reconstruction process provides
us with useful information about the resolution which we can expect from the
final reconstruction.
In this paper we make use of the so-called adjoint technique for calculat-
ing sensitivity functions. In the more traditional ‘direct method’ for calculating
sensitivities, typically as many forward problems need to be solved as there
are parameters to be reconstructed. In large scale inverse problems these are
typically several thousand parameters, one for each pixel or voxel used for
the forward modelling. It has been pointed out in the literature that in such
large scale parameter estimation problems the adjoint method can be much less
computationally expensive than this ‘direct method’ for calculating sensitivi-
ties [Ar95a, Ar95b, AP98, Dor00, FO96, HAO00, MO90, MOEH, Nat96]. This is
in particular true in cases where the Green’s functions of the underlying for-
ward problem (for example Maxwell’s equations) are not known analytically,
40 O. Dorn et al.
xsc xsc
xsc xsc xd
xs xd xs
xsc xsc
the rows (or columns) of the Jacobian or sensitivity matrix. They play a major
role in basically all gradient-based inversion methods which try to minimize
the least squares measure of the misfit between given data and predicted data
corresponding to a reference medium. Therefore, the ideas presented here are
certainly applicable also beyond the single step adjoint field inversion method
which we are concentrating on in this paper.
with
the magnetic field values at the receiver positions coincide with the data.
If there is no noise in the data, then there may be many parameter distri-
butions b̃ for (9), (10) that satisfy the data, so we have to select one using a
regularization criterion. If there is noise, the usual case, then we need to min-
imize the sum of the squares of the differences between the data calculated by
(9), (10) and those which are physically measured.
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 43
Notice that in the case of isotropic media we have b(x) = b(y) = b(z) = b such
that we typically will identify in that situation F = L2 (Ω) with b ∈ F . It will
always be clear from the context which space we refer to, such that we omit
using different symbols for this space in the cases of isotropic and anisotropic
media. All these quantities carry a physical dimension, which has to be taken
into account when introducing the following inner products for these spaces.
E1 E2 2 2
, = γE E1 E2 dx + γH H1 H2 dx,
H1 H2 U Ω Ω
J1 J2 2 2
, = γJ J1 J2 dx + γM M1 M2 dx,
M1 M2 Y Ω Ω
N N
(i) (i)
ζ1 , ζ2 Z = 2
γH h1,n h2,n = 2
γH h1,n h2,n
n=1 n=1 i=x,y,z
An analogous expression holds for the inner product of F = L2 (Ω) in the case
of isotropic media. The constants γE , γH , γJ , γM and γb are introduced in
order to make the given scalar products dimensionless. More precisely, we have
γE = [E]−1 , γH = [H]−1 , γJ = [J]−1 , γM = [M]−1 , and γb = [b]−1 , where the
44 O. Dorn et al.
symbol [.] denotes the dimension of the corresponding physical quantity. Notice
that γb γE = γJ , which follows immediately from (6). We will also consider in
this paper the corresponding dimensionless quantities which we always indicate
by a roof on top of the symbol, for example b̂ = γb b for b ∈ F or ĥ = γH h for
h ∈ Z. For these quantities, we will accordingly use the inner products , F̂ ,
, Ẑ , . . . , which are defined as above but without the additional factors in the
integrals. The spaces F , U , Y and Z (as well as F̂ , Û , Ŷ and Ẑ) are Hilbert
spaces with these inner products.
We write (6), (7) in system form
ΛM (b) u = q, (11)
with
⎛ ⎞
−b ∇×
⎜ ⎟
⎜ ⎟
ΛM (b) = ⎜ ⎟, (12)
⎝ ⎠
∇× a
and
E J
u = , q = .
H M
The Maxwell Operator ΛM (b) is defined on a dense subset of the Hilbert space
U, given the parameter distribution b. When the conductivity is everywhere
positive, σ > 0, (11) has a unique solution u ∈ U for each source q ∈ Y . For a
function u ∈ U of (11) we define the measurement operator M : U → Z by
(M u)n (x) = δ(x − dn ) H(x) dx (n = 1, . . . , N ) (13)
Ω
E
with u = H , and with dn , n = 1, . . . , N , being the detector positions. The
measurement operator M is well defined by (13) and maps the fields u ∈ U to
a vector of complex numbers M u ∈ Z, which consists of point measurements of
the magnetic field components at the detector positions dn , n = 1, . . . , N . More
precisely, we can write
E T
M = ( H(d1 ) , H(d2 ) , . . . , H(dN ) ) ,
H
where the symbol T stands for transpose. For the given source q and given data
g ∈ Z we define the residual operator R by
R : F → Z, R(b) = M u(b) − g,
where u(b) is the solution of (11) with parameter b. The operator R is a nonlinear
operator from the space of parameters F to the space of measurements Z.
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 45
ms = α1 e1 + α2 e2 + α3 e3 , (14)
with ΛM (b) defined as in (12), and a, b defined as in (5). In order to find a useful
representation of the linearized residual operator R of R, we perturb b, E and
H by
b → b + δb E →E+v H → H + w,
and plug this into (6), (7). When neglecting terms of higher than linear order
in δb, v and w, and applying the measurement operator M , we arrive at the
following result.
Theorem 4.1. The linearized residual operator R [b] is given by
v
R [b] δb = M , (16)
w
v
where w solves
v
δb E
ΛM (b) = . (17)
w 0
It is a mapping from the data space Z into the parameter space F . In the next
theorem, we will present a useful expression for the action of R [b]∗ on a given
vector ζ ∈ Z. The proof of this theorem can be found in [DBBP2].
Theorem 5.1. Let us assume that we are given ζ = (h1 , . . . , hN ) ∈ Z with
(x) (y) (z)
ĥn = (ĥn , ĥn , ĥn )T ∈ C3 for n = 1, . . . , N . Furthermore, we denote by
E(x) = (E (x), E (x), E (z) (x))T the solution of the adjoint problem
(x) (y)
46 O. Dorn et al.
E 0
Λ∗M (b) = −1 N (i) , (18)
H γM n=1 i=x,y,z ĥn ei δ(x − dn )
with ⎛ ⎞
−b ∇×
Λ∗M (b) = ⎝ ⎠, (19)
∇× a
b = −iω + σ, a = −iωµ.
∗
Then R [b] ζ ∈ F is given by
R [b]∗ ζ (x) = γb−1 diag Ê (x) (x)Ê (x) (x) , Ê (y) (x)Ê (y) (x) , Ê (z) (x)Ê (z) (x) ,
(20)
where E is the solution of (15).
In this section, we will present the basic decompositions of the linearized residual
operator and its adjoint into sensitivity functions. The proofs of the theorems in
this section and in the next section can be found in [DBBP2].
(i)
Theorem 5.2. Given a source (14). We can find functions Sk (n, l; y), i =
1, 2, 3, such that the linearized residual operator R [b] can be decomposed as
3
3 3
R [b]δb = γ −1 ! (i) (y)
δb
(i)
α̂k S (n, l; y) dy el . (21)
H k
n Ω
l=1 i=1 k=1
(i)
Moreover, the same functions Sk (n, l; y) also decompose the adjoint linearized
residual operator R [b]∗ in the following way
3
(i) N 3
∗ −1 (l) (i)
R [b] ζ (y) = γb ĥn α̂k Sk (n, l; y) . (22)
n=1 l=1 k=1
(i)
The functions Sk (n, l; y) are called anisotropic sensitivity functions.
(i)
In the following, we will first define these sensitivity functions Sk (n, l; y),
and then reformulate the above theorem in a slightly more formal way. First, we
decompose the given vector ζ = (h1 , . . . , hN ) ∈ Z as
where the fields (E, H)T are solutions of (15). Notice that these sensitivity func-
tions are defined to be dimensionless.
Now, making use of the inner products which have been introduced in Sect. 4,
we get the following equivalent formulation of Theorem 5.2.
(i)
Theorem 5.3. Given a source (14). Define the functions Sk (n, l; y) as in (26).
Then we have the following decompositions of the linearized residual operator
R [b] and its adjoint R [b]∗ .
Projection Step:
3 3
3
R [b]δb = γ −1 ! (i) , S (i) (n, l; . )
α̂k δb el . (27)
H k
n F̂
l=1 i=1 k=1
Backprojection Step:
(i) 3
∗
γb−1
(i)
R [b] ζ (y) = α̂k ζ̂ , Sk ( . , . ; y) . (28)
Ẑ
k=1
Remark 5.1. We note that (27) gives rise to the interpretation of R [b] as a ‘pro-
jection operator’. It ‘projects’ each component δb(i) of the parameter perturba-
(i)
tion δb along the sensitivity function Sk (n, l; . ) into the data space Z. Summing
up all these contributions yields the (linearized) residuals R [b]δb. Analogously,
(28) gives rise to the interpretation of R [b]∗ as a ‘backprojection operator’. It
‘backprojects’ the residual vector ζ of the data space Z along the sensitivity
(i)
functions Sk ( . , . ; y) into the parameter space F .
Let us consider an isotropic medium, i.e. let biso (y), δbiso (y) be given by
biso (y) = diag b(0) (y), b(0) (y), b(0) (y) ,
δbiso (y) = diag δb(0) (y), δb(0) (y), δb(0) (y) .
48 O. Dorn et al.
We can replace the diagonal tensors for describing the parameters δbiso and
biso by the scalar functions δb(0) , b(0) , which live in the reduced function space
L2 (R3 ). The isotropic sensitivity functions Sk (n, l; y) are defined as
Backprojection step:
3
R [b]∗ ζ (y) = γb−1 α̂k ζ̂ , Sk ( . , . ; y) . (34)
Ẑ
k=1
Rj (b̃) = 0, j = 1, . . . , p. (35)
The index j refers here to the actual source position qj which is considered by
the operator Rj (b̃). Using real data, we cannot be sure whether these equations
can be satisfied exactly. Therefore, we generalize our criterion for a solution.
Defining the least squares cost functionals
1
Jj (b) = Rj (b)2L2 (36)
2
for j = 1, . . . , p, we are searching for a joint minimizer of these cost functionals.
A standard method for finding a minimizer of (36) is to start a search with some
initial guess b(0) , and to find descent directions for the sum
p
J (b) = Jj (b)
j=1
in each step of an iterative scheme. Popular choices for descent directions are
for example the gradient direction, conjugate gradient directions, Newton- or
Quasi-Newton directions (see for example [DS96, NW99, Vog02] for details). In
our Kaczmarz-type approach, however, we consider only individual terms of (36)
in a given step. We search for descent directions only with respect to the chosen
term Jj (b), and apply it to the latest best guess. In the next step, we will then
choose a different term Jj (b) of (36) and proceed in the same way. After having
used each of the terms in Jj (b) exactly once, we have completed one sweep of
our algorithm.
This strategy of switching back and forth between different data subsets
does not only have the advantage that a correction of our latest best guess
for the parameters can be achieved more rapidly, but it also helps to avoid
getting trapped in so-called ‘local minima’ of more classical gradient-based search
algorithms. This is because the variety of search criteria is increased in our
algorithm. Instead of having just one search direction available per data set
(considering only descent directions of J (b)), we have now as many useful search
directions as we have source positions. Even if one of these search directions does
not advance our reconstruction, typically at least one of the remaining search
directions gives us an improvement of it. Notice also that the combined cost
J (b) is certainly allowed to increase for a while during the reconstruction using
the Kaczmarz-type scheme. This will for example happen when a gradient-based
inversion algorithm with respect to J (b) would get close to a local minimum,
50 O. Dorn et al.
but at least one of the individual Kaczmarz-type search directions is not affected
by this local minimum of the combined least-squares cost functional.
We will in the following derive our Kaczmarz-type update directions for re-
ducing the individual terms (36) using the formulation (35). These update direc-
tions will have the useful property that they can easily be generalized in order
to incorporate efficient regularization schemes in our algorithm.
In order to find an ‘update’ (or ‘correction’) δb for our parameter b we lin-
earize the nonlinear operator Rj (assuming that this linearized operator Rj [b]
exists and is well-defined) and write
The linearized operator Rj [b] is often called the Fréchet derivative of Rj at b.
(See for example [DDNPS] and references therein for some formal derivations
of Fréchet derivatives in different applications.) It is also closely related to the
‘sensitivity functions’ of the parameter profile with respect to the data. Using
(37) we want to look for a correction δb such that Rj (b + δb) = 0. Neglecting
terms of order O(||δb||2 ) in (37), this amounts to solving
where Rj [b]∗ is the adjoint operator of Rj [b] with respect to our chosen spaces
F and Z. In applications with very few data, this form has the useful property
that it avoids contributions in the solution which are in the (often non-empty)
null-space of the (linearized) forward operator Rj [b]. Using (37) it can be verified
by direct calculation that
for a step-size κ > 0, such that (39) also is a descent direction of the least squares
cost functional (36).
In our application the operator C = (Rj [b]Rj [b]∗ )−1 is very ill-conditioned,
such that a regularized version needs to be used. This can be, for example,
Ĉ = (Rj [b]Rj [b]∗ + λI)−1 where λ is some regularization parameter and I is the
identity operator. Unfortunately, in practice both, C and Ĉ, are very expensive
to calculate and to apply to the residuals Rj . Typically, a direct calculation of
the operator Ĉ would require us to solve as many forward and adjoint problems
as we have independent data values.
When using a very large regularization parameter λ, the contribution of
Rj [b]Rj [b]∗ can be neglected and we end up with essentially (i.e. up to the
scaling factor λ−1 ) calculating
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 51
In this section we want to show how the adjoint residual operator is applied in
the case where many receiver positions are used simultaneously for a given source
and where the medium is assumed to be possibly inhomogeneous but isotropic.
A formal derivation of this result can be found in [DBBP1].
Let Ĉj denote some suitably chosen approximation to the operator Cj , which
might depend on the source index j. For the inversion, we will need some efficient
way to compute
Rj (b)∗ ζj
52 O. Dorn et al.
be the data vector, which in the given step j of the inversion scheme usually con-
sists of the (filtered) residual values of all receiver positions which correspond to
the source qj . We will in the following omit all dimensional factors (for example
γH ) to simplify the notation. Then Rj (b)∗ ζj ∈ F is given by
" ∗ #
Rj (b) ζj (x) = Ej (x) · Ej (x), (46)
with ⎛ ⎞ ⎛ ⎞
−b ∇× −b̄ ∇×
ΛM (b) = ⎝ ⎠, Λ∗M (b) = ⎝ ⎠, (49)
∇× a ∇× ā
b = iω + σ, b̄ = −iω + σ,
(50)
a = iωµ, ā = −iωµ.
We see that we only have to solve one forward problem (47) and one adjoint
problem (48) in order to calculate Rj (b)∗ ζj in an efficient way (compare Fig. 2).
We will now describe in detail the algorithm which we use in the numerical
experiments. In brief algorithmic form, the iterative scheme can be written as
(0)
bp = b(0)
sweep loop: DO i = 1, Imax
(i) (i−1)
b0 = b p
source loop: DO j = 1, p
= −Rj (bj−1 )∗ Ĉj Rj (bj−1 )
(i) (i) (i)
δbj
(i) (i) (i)
bj = bj−1 + κj δbj
END DO source loop
END DO sweep loop
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 53
Here, b(0) is some initial guess for b and Imax is the total number of sweeps.
The parameter κj is the step-size in step j, and Ĉj the chosen approximation to
Cj . For example, b(0) can be taken to be the correct layered background parame-
ter distribution. The aim will then be to detect some unknown inclusions which
are imbedded in this layered background medium, see for example [DBBP1].
Notice that in one single step of the algorithm we have to solve one forward
(i)
problem (47) to compute Rj (bj−1 ), and one adjoint problem (48) to compute
(i) (i)
δbj from a given Rj (bj−1 ).
We mention that, in the algorithm presented so far, no explicit regularization
has been applied. If there is a sufficient amount of independent data available,
this algorithm already yields good reconstructions. However, for practical ap-
plications it would be desirable to have an efficient toolbox of regularization
schemes available which can be used for stabilizing the inversion in situations
with only very few data, and in order to incorporate prior information into the
reconstructions. In the following section, we propose such an efficient and flexi-
ble regularization tool. It has been tested and evaluated numerically with great
success in a different but related geophysical application in [GKMD]. The nu-
merical evaluation of this scheme in the situation of 3D EMIT still needs to be
done, and is planned as part of our future research.
For reasons explained below, we will instead prefer to work with the equivalent
norm 1/2
mα,β := αm2L2 + β∇m2L2 , α, β > 0
and its associated inner product
A proper choice of the weighting parameters α and β will allow us to steer the
regularization properties of our algorithm in an efficient and predictable way.
Let us denote the new parameter space H1 (Ω), when equipped with the
weighted norm . α,β , by F̂ . When using this modified space in our algorithm,
we also have to adjust the operators acting on it, in particular the adjoint of
the linearized residual operator. This operator is now required to map from the
data space Z into F̂ . Moreover, the minimum norm solution of (38) is now taken
with respect to the weighted norm . α,β , which clearly gives us a different
candidate. The necessary adjustments for our algorithm can be done as follows.
Denote as before by Rj [b]∗ ζ the image of ζ ∈ Z under application of the
adjoint linearized residual operator as calculated in Sect. 6.2, considered as an
operator mapping from Z into F . Denote furthermore by Rj [b]◦ ζ its image un-
der the adjoint linearized residual operator with respect to the newly defined
weighted inner product, mapping into the smaller space F̂ . With a straightfor-
ward calculation, using the definitions of the two adjoint operators
Rj [b]x, ζZ = x, Rj [b]∗ ζF = x, Rj [b]◦ ζF̂ , (51)
it follows that
−1
Rj [b]◦ ζ = (αI − β∆) Rj [b]∗ ζ, (52)
−1
where we supplement the inverted differential operator (αI − β∆) by the
boundary condition ∇(Rj [b]◦ ζ) · n = 0 on ∂Ω. The symbol I stands for the
identity, and ∆ stands for the Laplacian operator. Equation (52) can be easily
derived by applying Green’s formula to the right hand side equality in (51).
In practice, the ratio γ = β/α (which can be considered being a ‘regular-
ization parameter’) is an indicator for the ‘smoothing properties’ of our scheme.
The larger this ratio, the more weight is put on minimizing the derivatives of our
solution. Therefore, by properly choosing this ratio, we can steer the smoothness
properties of our final reconstruction to a certain degree. In our numerical ex-
periments, we will choose this ratio once, when starting the algorithm, and keep
it fixed during the iterations. The other free parameter, say α, will be chosen in
each individual step to scale the update properly. In our numerical experiments,
we choose α such that
simply putting it to 1 right at the beginning. When testing and comparing the
performance of different regularization parameters γ it is practically useful that
the order of magnitude of the calculated values of Rj [b]◦ ζ does not depend too
much on γ.
Notice also that the new search directions using this modified adjoint operator
are still descent directions for the least squares cost functional (36), as can be
verified easily by replacing F by F̂ in (40) and (42).
Practically, the scheme is implemented as follows:
Fig. 3. Shown are different cross-sections through the real part of the numerically
calculated isotropic sensitivity function S(z, z; y) as a function of y for an infinitely
extended homogeneous medium. The source and the receiver are z-directed magnetic
dipoles at depth 80 m. The scaling factor is Rmax = 3.62 × 10−13
Fig. 4. Shown are the same cross-sections through the real part of the isotropic sensitiv-
ity function S(z, z; y) as in Fig. 3, but now calculated analytically using the expressions
derived in [DBBP2]. The scaling factor is again Rmax = 3.62 × 10−13 . The agreement
with Fig. 3 is very good
58 O. Dorn et al.
Fig. 5. Shown are different cross-sections through the imaginary part of the numerically
calculated isotropic sensitivity function S(z, z; y) as a function of y for an infinitely
extended homogeneous medium. The source and the receiver are z-directed magnetic
dipoles at depth 80 m. The scaling factor is Imax = 2.96 × 10−12
notation which indicates that we consider only one receiver.) In this simplified
notation, the first argument z indicates the orientation of the dipole source,
and the second argument z says that we are measuring the z-component of the
magnetic field. The argument y is the independent variable of the sensitivity
function ranging over all possible scattering positions in the medium.
We mention that all sensitivity functions in this paper are displayed on a
logarithmic scale in the following sense. Let us define the values Rmax and Imax
according to
$ $
$ $
Rmax = max $Re(S(z, z; y))$ ,
y∈Ω
$ $
$ $
Imax = max $Im(S(z, z; y))$ ,
y∈Ω
and analogous expressions for S(z, y; y). Here, the maximum is taken over the
finite number of computer evaluations of S(z, z; y) which is given by the FDFD
forward modelling code. This has the effect that all values of S(z, z; y) which
contribute to Rmax and Imax are actually finite, whereas the theoretically cal-
culated values would become singular at the source and receiver positions. With
these reference values Rmax and Imax , we define the logarithmic sensitivity func-
tions by
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 59
Fig. 6. Shown are different cross-sections through the real part of the numerically
calculated isotropic sensitivity function S(z, y; y) as a function of y for an infinitely
extended homogeneous medium. The source is a z-directed magnetic dipole at depth
80 m, and the receiver is a y-directed magnetic dipole at depth 80 m. The scaling factor
is Rmax = 3.36 × 10−13
Rmax
logreal(z, z; y) = −sign(Re(S(z, z; y))) 60 − 10 × log
|Re(S(z, z; y))| +
Imax
logimag(z, z; y) = −sign(Im(S(z, z; y))) 60 − 10 × log ,
|Im(S(z, z; y))| +
and analogous expressions hold for S(z, y; y). Here, sign(x) gives the sign (‘plus’
or ‘minus’) of the real number x, the symbol (x)+ = max{0, x} acts as the
identity for positive values of x and delivers zero for negative values. ‘Re’ means
‘real part’, and ‘Im’ means ‘imaginary part’. It is these logarithmic sensitivity
functions which are displayed in Figs. 3–9.
In Fig. 4 we have displayed the same sensitivity functions as in Fig. 3, but
now being calculated with an independent analytical expressions which has been
derived in [DBBP2]. From the agreement of the two Figs. 3 and 4 we can con-
clude that our numerically calculated sensitivity functions are indeed correct.
Additional numerical verifications have been done in [DBBP2]. Notice also that
all our sensitivity functions have been evaluated on the grid positions of our
staggered FDFD grid.
Figure 5 displays the imaginary part of the sensitivity function for the same
setup as in Figs. 3 and 4. In Figs. 6 and 7 the real and imaginary part of the
60 O. Dorn et al.
Fig. 7. Shown are different cross-sections through the imaginary part of the numerically
calculated isotropic sensitivity function S(z, y; y) as a function of y for an infinitely
extended homogeneous medium. The source is a z-directed magnetic dipole at depth
80 m, and the receiver is a y-directed magnetic dipole at depth 80 m. The scaling factor
is Imax = 2.96 × 10−12
We have presented in Sect. 8.1 some results of our numerical experiments for
calculating sensitivity functions. In the following, we want to indicate a few
problems which come up in typical imaging scenarios, and where the study of
sensitivity functions might be very useful in order to address these problems
successfully.
Adjoint Fields and Sensitivities for 3D Electromagnetic Imaging 61
Fig. 8. Shown are different cross-sections through the real part of the numerically
calculated isotropic sensitivity function S(z, y; y) as a function of y for an infinitely
extended homogeneous medium. The source is a z-directed magnetic dipole at depth
80 m, and the receiver is a z-directed magnetic dipole at depth 7 m. The scaling factor
is Rmax = 4.00 × 10−13
First, when imaging a geophysical test site, some decision has to be made
about which degree of anisotropy should be expected in the given situation.
For a discussion of some implications of electrical anisotropy in applications
see also Weidelt [Wei95]. It can be observed in our numerical experiments (not
shown here, but see [DBBP2]) that the shapes of the x, y and z-component
sensitivity functions typically differ from each other, and certain source-receiver
constellations are very sensitive to some of these components and much less
sensitive to others. This indicates that it should in general be possible to resolve
anisotropic structures in the earth from electromagnetic data, provided that a
sufficient amount of appropriately gathered data is available. On the other hand,
it indicates that the treatment of strongly anisotropic media with an isotropic
imaging approach can lead to inaccurate reconstructions because of the related
misinterpretations of the data.
Second, in basically all imaging problems the question comes up which degree
of resolution can be expected from the reconstructions. In our situation, since
the reconstructions are essentially results of a series of backprojection steps, only
features of the medium can be recovered which are actually represented by the
sensitivity functions. This gives us some information about the resolution which
we can expect. If we want to further quantify the degree of resolution, we could
use for example the concept of a ‘resolution matrix’ which has been discussed
62 O. Dorn et al.
Fig. 9. Shown are different cross-sections through the imaginary part of the numerically
calculated isotropic sensitivity function S(z, y; y) as a function of y for an infinitely
extended homogeneous medium. The source is a z-directed magnetic dipole at depth
80 m, and the receiver is a z-directed magnetic dipole at depth 7 m. The scaling factor
is Imax = 1.22 × 10−12
Acknowledgments
Part of the work presented here has been performed during the research visit of
O.D. at UCLA, Los Angeles, in the fall of 2003 for the IPAM special program
on ‘imaging’. The support of IPAM is gratefully acknowledged.
References
[AM95a] Alumbaugh D L and Morrison H F 1995 Monitoring subsurface
changes over time with cross-well electromagnetic tomography Geophys-
ical Prospecting 43 873–902
[AM95b] Alumbaugh D L and Morrison H F 1995 Theoretical and practical con-
siderations for crosswell electromagnetic tomography assuming cylindrical
geometry Geophysics 60 846–870
[AB01] Ammari H and Bao G 2001 Analysis of the scattering map of a linearized
inverse medium problem for electromagnetic waves Inverse Problems 17
219–234
[Ar95a] Arridge S R 1995 Photon-measurement density functions. Part I: Analyt-
ical forms Applied Optics 34 (34) 7395-7409
[Ar95b] Arridge S R and Schweiger M 1995 Photon-measurement density func-
tions. Part II: Finite-element-method calculations Applied Optics 34 (31)
8026-8037
[Ar99] Arridge S (1999) Optical Tomography in Medical Imaging Inverse Prob-
lems 15 R41-R93
[AP98] Ascher U M and Petzold L R (1998) Computer Methods for Ordi-
nary Differential Equations and Differential-Algebraic Equations (SIAM:
Philadelphia)
[BK90] Berryman, J. G., and R. V. Kohn, 1990, Variational constraints for elec-
trical impedance tomography, Phys. Rev. Lett. 65, 325–328
[BBP96] Borcea L, Berryman J G, and Papanicolaou G C 1996 High Contrast Im-
pedance Tomography, Inverse Problems 12 935–958
[BBP99] Borcea L, Berryman J G, and Papanicolaou G C 1999 Matching pursuit
for imaging high contrast conductivity Inverse Problems 15 811–849
[Bor02] L. Borcea 2002 Electrical Impedance Tomography, Inverse Problems 18
R99–R136
[BB99] Buettner H M and Berryman J G 1999 An electromagnetic induction
tomography field experiment at Lost Hills, CA, in the Proceedings of
SAGEEP, Oakland, CA, March 14–18, 663–672
[CBB01] N. J. Champagne II, J. G. Berryman, and H. M Buettner, FDFD (2001) A
3D finite-difference frequency-domain code for electromagnetic induction
tomography J. Comput. Phys. 170 830–848
[Cla76] Claerbout, J. F., 1976, Fundamentals of Geophysical Data Processing: With
Applications to Petroleum Prospecting, McGraw-Hill, New York.
[DRLN] Daily, W. D., A. Ramirez, D. LaBrecque, and J. Nitao, 1992, Electrical
resistivity tomography of vadose water movement, Water Resources Res.
28, 1429–1442
[DA89] Davis, J. L., and A. P. Annan, 1989, Ground penetrating radar for high
resolution mapping of soil and rock stratigraphy, Geophysical Prospect. 37,
531–551
64 O. Dorn et al.
Miguel Moscoso
Summary. In this study I present polarization effects resulting from the reflection
and transmission of a narrow beam of light through biological tissues. This is done
numerically with a Monte Carlo method based on a transport equation which takes into
account polarization of light. We will show both time-independent and time-dependent
computations, and we will discuss how polarization can be used in order to obtain
better images. We consider biological tissues that can be modeled as continuous media
varying randomly in space, containing inhomogeneities with no sharp boundaries.
1 Introduction
with a constant net flux [8]. When the incident laser beam is narrow no solution
is known. Therefore we have developed a Monte Carlo method to obtain the
spatial distribution of the total intensity and of the polarization components
of the transmitted and reflected beams. Other methods for solving the vector
radiative transfer equation have been explored in [9].
In [9] we presented a complete discussion of Chebyshev spectral methods
for solving radiative transfer problems. In this method, we approximated the
spatial dependence of the intensity by an expansion of Chebyshev polynomials.
This yields a coupled system of integro-differential equations for the expansion
coefficients that depend on angle and time. Next, we approximated the integral
operation on the angle variables using a Gaussian quadrature rule resulting in
a coupled system of differential equations with respect to time. Using a second
order finite-difference approximation, we discretized the time variable. We solved
the resultant system of equations with an efficient algorithm that makes Cheby-
shev spectral methods competitive with other methods for radiative transfer
equations.
Section 2 contains the theory of radiative transfer employing Stokes para-
meters. In Sect. 3 we describe our model where we represent a biological tis-
sue as a medium with weak random fluctuations of the dielectric permittivity
(r) = 0 [1 + δ(r)], where r denotes position and δ(r) is the continuous ran-
dom fractional permittivity fluctuation. In Sect. 4 we describe a Monte Carlo
method for three-dimensional problems which take into account the vector na-
ture of light. Section 5 describes our numerical calculations. Section 6 contains
our conclusions.
er
θ
el
y
ϕ
x
Fig. 1. Coordinate system for the Stokes vector. The shaded plane φ = contant con-
taining the wavevector k is called the meridian plane. We choose el to lie in this plane
and to be orthogonal to k
where
I = El El∗ + Er Er∗ ,
Q = El El∗ − Er Er∗ ,
U = El Er∗ + Er El∗ ,
V = iEl Er∗ − Er El∗ . (2)
Here El and Er are the complex amplitudes of the electric field referred to an
orthonormal system (k̂, el , er ); see Fig. 1. I is the total intensity. The Stokes
parameter Q is the difference between the intensities transmitted through two
linear analyzers in orthogonal directions. U is defined like Q, but with reference
to linear analyzers rotated by 45◦ with respect to those of Q. V is the difference
between the intensities passing through right and left circular analyzers.
The Stokes vector I(x, k̂, t), defined for all directions k̂ at each point x and
time t, satisfies the transport equation [8]
1 ∂I(x, k̂, t)
+ k̂ · ∇x I(x, k̂, t) + (Σ (s) (x) + Σ (a) (x))I(x, k̂, t) =
v ∂t
Σ (s) (x) F (k̂ · k̂ )I(x, k̂ , t)dk̂ . (3)
Here v, Σ (s) , and Σ (a) are the velocity of the light in the medium, the scattering
coefficient, and the absorption coefficient, respectively. The 4 × 4 scattering ma-
trix F (k̂ · k̂ ) in (3) describes the probability for a photon entering a scattering
70 M. Moscoso
process with Stokes vector I(i) and direction k̂ to leave the process with Stokes
vector I(s) and direction k̂. The dot product in the argument of F means that
it depends only on the cosine of the scattering angle Θ.
3 Model
We consider biological tissue with weak random fluctuations of the dielectric
permittivity (r) = 0 [1 + δ(r)]. Then F is related to the power spectral density
R̂(q) of δ(r) by [12]:
π 4 Θ
F (Θ) = k R̂(2k sin )S(Θ) . (4)
2 2
Here ⎛ ⎞
1 + cos2 Θ cos2 Θ − 1 0 0
⎜ ⎟
⎜ 2 ⎟
1⎜
⎜
cos Θ − 1 1 + cos2 Θ 0 0 ⎟
⎟
S(Θ) = ⎜ ⎟ (5)
2⎜ 0 0 2 cos Θ 0 ⎟
⎝ ⎠
0 0 0 2 cos Θ
ε2 l 3
R̂(k) = . (7)
π 2 (1 + k 2 l2 )2
Polarization-Based Optical Imaging 71
The exponential correlation function (6) is a especial case of the von Karman
correlation function
ε2 21−κ r κ r
R(r) = ( ) Kκ ( ) , κ ∈ [0, 0.5] (8)
Γ (κ) l l
Monte Carlo methods are based on the stochastic nature of the propagation of
light in a random medium. They simulate photon histories according to funda-
mental probability laws which depend on the scattering medium. Their main
advantage is their relative simplicity and their ability to handle complicated
geometries. They also provide exact solutions (up to statistical errors) of the ra-
diative transport equation (3). Their main drawback is that they require a large
number of photons to obtain statistical accuracy for large optical depths. The
rate of convergence is given by the central limit theorem of probability theory. It
states that for N independent random variables I i from theNsame distribution,
with finite mean m and variance σ 2 , the average IN = i I i /N is normally
distributed about m with variance σ 2 /N . This means that IN converges to the
72 M. Moscoso
√
expected value at the rate const./ N . To make it converge faster several vari-
ance reduction techniques can be used [7].
We use two basic approaches to improve the convergence [17, 18]. First, we
modify the random walk sampling to follow more particles in important regions
where the scoring contributes more to the final tally. This is especially relevant
in computing light transmission through a medium of large optical thickness. In
that case, only a small fraction of particles penetrates into the deeper regions
of the medium, producing bad statistics in the computation. Second, we modify
the scoring method to maximize the statistical information of a given random
walk. In doing this, a trajectory is no longer associated with a single particle.
The replacement of the analog simulation by a non-analog one is usually called
a variance reduction technique.
In the next subsection we describe the simplest case of the analog simula-
tion. In Sect. 4.2 we explain the geometry splitting and Russian roulette, whose
objective is to spend more time sampling particles in important regions and less
time sampling particles in unimportant regions. In Sect. 4.3 we explain the point
detector technique, which is a deterministic estimate of the flux at a point from
each collision.
with k̂in = (sin θin cos φin , sin θin sin φin , cos θin ). Next, the scattering direction
k̂f is determined according to the probability density function [7, 19]
1 % &
P (k̂f , k̂in , Qin , Uin ) = F̄11 + F̄12 Qin + F̄13 Uin . (12)
Σs
Here F̄ij , with i, j = 1, . . . , 4, represents the (i, j) element of the scattering ma-
trix F̄ referred to the meridian plane (plane of constant φd ) as plane of reference
Polarization-Based Optical Imaging 73
for the Stokes vectors. See [8] for the required rotation. P depends on the di-
rection of incidence and scattering, and on the incoming polarization state. The
dependence on the polarization state is essential because linearly polarized pho-
tons cannot be scattered in the direction of oscillation of the electric field. This
property, related to the vector nature of light, is lost if the probability density
function does not depend on the polarization. Note that P does not depend on
Vin because F14 = 0 in weakly fluctuating media. P has been normalized so that
for all incident directions and polarization states
P (k̂f , k̂in , Qin , Uin )dk̂f = 1 , (13)
Equation (14) does not depend on R̂(k). The normalization in (14) assures that
the intensity of the scattered photon remains equal to the incident one. From
(14) it also follows that if the incident photon is completely polarized then the
scattered photon is also completely polarized.
Next, a new path length is calculated from (10) and the tracing continues
using (11)–(14) until the photon exit the medium.
If a photon reaches the detector at an angle with the normal to the sample
surface that is less than its semiaperture, then its contribution to the quantities of
interest (intensity and polarization) are stored. If the incident beam is circularly
polarized, two quantities are computed: the total intensity I and the circular
component V . In the case of a linearly polarized incident beam, three intensities
are computed: the total intensity I, the parallel-polarized component Il , and
the cross-polarized component Ir . If a photon of intensity dI = 1 arrives at the
detector with polar direction (θd , φd ), these two last intensities can be calculated
easily from the Stokes parameters Qd and Ud of the photon. The direction of
polarization makes an angle
with the meridian plane. The sign is positive for transmission, and negative for
reflection. The difference in signs is due to the change of sign of U when the z
component of the wave-vector k̂ changes sign. We have chosen el × er = k̂ (see
Fig. 1). If el × er = −k̂ instead, then the signs must be changed in (15). The
contribution of the photon to the transmitted intensities is:
I = dI , (16)
Il = dI[cos2 (φd + χd ) cos2 θd + sin2 θd ] , (17)
Ir = dI[sin2 (φd + χd ) cos2 θd + sin2 θd ] . (18)
74 M. Moscoso
To compute the reflected intensities, we must change χd to −χd in (17) and (18).
(n) (n)
After N photons have been traced, and the intensities I (n) , Il , Ir , and
(n)
V of the N histories have been stored, we average to obtain the statistical
result,
N N
1 1 (n)
I¯ = I (n) , I¯l = Il , (19)
N n=1
N n=1
N N
1 1
I¯r = Ir(n) , V̄ = V (n) . (20)
N n=1
N n=1
We note from (17) and (18) that Il + Ir = I(1 + sin2 θd ). This is not equal to
I unless sin θd = 0, because if sin θd = 0 there is a component of the electric field
normal to the analyzers. This component is transmitted through them indepen-
dently of their orientation. Consequently, when measuring a bundle of light, in a
typical experimental situation, the sum I¯l + I¯r of the parallel and cross-polarized
components is not equal to the total intensity I, ¯ measured without polarization
analyzers.
We also note that the fraction of light with the original polarization, (I¯l −
¯r )/I,
I' ¯ coincides with the common definition of the degree of polarization
¯ only for small apertures. In this case, the intensity of light that
Q̄ + Ū 2 /I,
2
has maintained the incident polarization state, I¯l − I¯r , is a measure of the Stokes
parameter Q̄. In the problem that we shall consider, Ū is zero due to symmetry.
One of the most widely used variance reduction methods is the combination of
the so called importance splitting and Russian roulette techniques. The problem
domain is divided into M subdomains with different importances I (m) (m =
0, . . . , M − 1). We will assume for simplicity that I (m) = cm where c is a chosen
positive integer. The more probable it is that a particle in a sub-domain can score
at the detector, the more important is the sub-domain. Let a particle of weight
W evolve between two collisions i and i + 1, and let us denote the importance
of the particle by Ii and Ii+1 at each collision (Ii = I (m) if at collision i the
photon is in sub-domain m). We form the importance ratio η = Ii+1 /Ii . If η = 1
the importance of the photon does not change and its transport continues in the
usual way. If η > 1 the photon has moved to a more important region and it
is split into Nsplit = η photons. The weight W of each of the split photons is
divided by Nsplit . If η < 1 it has entered into a less important region, and a
Russian roulette game is played where the particle has probability psurv = η
of surviving. A random number ξ is drawn uniformly on the interval (0, 1) and
compared with psurv . If ξ < psurv the photon survives and its weight is increased
by 1/psurv . Otherwise, it is eliminated. In doing this, however, a balance has to
be struck. Implementing splitting generally decreases the history variance but
increases the computing time per history. On the contrary, Russian roulette
Polarization-Based Optical Imaging 75
increases the former and decreases the latter, so that more histories can be run.
The usual measure of efficiency of the Monte Carlo simulations is the figure of
merit 1/σ 2 t, where the variance σ 2 measures the inaccuracy of the result, and
t is the mean time per history. We have seen experimentally that, for example,
for a medium of optical thickness 12, the choice c = 2 and M = 5 improves
the figure of merit by a factor 10. However, for optical thickness less than 8 we
have not found the splitting/Russian roulette necessary. For deep penetration,
experience has indicated that the best splitting results are achieved keeping the
particle flux through the medium close to a flat distribution, and that there is
a broad range of parameters close to the optimal choice [18]. If splitting with
Russian roulette is the only variance reduction technique used, all particles in
the same sub-domain m have the same weight 1/I (m) , assuming that all particle
histories began with weight 1.
We now consider scoring at a small detector of area dA, often called a point
detector. For each scattering event i of the n-th photon, we store the probability
(n)
that the n-th scattered photon, of weight Wi , will arrive at the detector with-
out further collisions. Let P (Θ, P)dΩ be the probability of scattering into the
solid angle dΩ about Θ, where Θ is the scattering angle directed to the detector,
and P = (Q0 , U0 ) denotes the liner polarization state of the photon. Taking into
account the attenuation suffered by particles traveling through the medium, the
probability of arriving at the detector is
(R
P (Θ, P)dΩe− 0
Σs (l)dl
, (21)
linearly polarized incident beam, the contributions to the parallel and cross-
polarized components are evaluated with expressions similar to (17) and (18),
respectively.
After N photons have been traced, the statistical averages are computed as
N Sn N Sn
1 (n) 1 (n)
F̄I = FIi , F̄l = Fli , (24)
N n=1 i=1
N n=1 i=1
N Sn N Sn
1 (n) 1 (n)
F̄r = Fri , F̄V = FV i , (25)
N n=1 i=1
N n=1 i=1
5 Numerical Results
receiver
incident beam
scan direction
x
y
s Z
L x
y d
z receiver
L
(s) (s)
Scattering coefficients of the background Σ bg , and of the inclusion Σ incl .
(a) (a)
Absorption coefficients of the background Σ bg , and of the inclusion Σ incl .
Fig. 2. The setup used in the simulations. The dimensions of the sample are L×L×Z. It
consists of a uniform medium in which a spherical object of radius 5.0 mm, with different
optical properties, is imbedded. The center of the object is at 1.6 cm from the input
surface. The distance between two consecutive incident beam positions is 2.5 mm. The
detectors are placed at a distance d = 1.0 cm from the exit surface. The semiaperture
angle of the each detector is 90◦ . A complete scan consists of 17 measurements
78 M. Moscoso
0.9 0.9
0.85 0.85
0.75 0.75
−2 −1 0 1 2 −2 −1 0 1 2
0.95 0.95
0.9 0.9
0.85 0.85
0.75 0.75
−2 −1 0 1 2 −2 −1 0 1 2
Distance (cm) Distance (cm)
Fig. 3. Reflection. Σbg = Σincl = 1 cm−1 , Σbg = 0.1 cm−1 , Σincl = 100 cm−1 . The
(s) (s) (a) (a)
solid lines show the total intensity I. The dashed lines show (a) the parallel-polarized
component Il , (b) the cross-polarized component Ir , (c) the polarization-difference
Il − Ir , and (d) the degree of polarization (Il − Ir )/I
the scattering matrix S is the same for inclusion and background. However the
scattering coefficients and the absorption coefficients may differ.
Figures 3 and 4 show numerical results for the time-integrated backscattered
(s) (s)
and transmitted light, respectively. The optical parameters are Σbg = Σincl =
1 cm−1 , Σbg = 0.1 cm−1 , Σincl = 100 cm−1 . The solid lines show the total
(a) (a)
intensity I, and the dashed lines show: (a) the parallel-polarized component
Il , (b) the cross-polarized component Ir , (c) the polarization-difference Il − Ir ,
and (d) the degree of polarization (Il − Ir )/I. We observe in Fig. 3 that the
backscattered cross-polarized component Ir gives the best image of the absorbing
object. The reason for the enhanced resolution is that this image is formed
by photons that probe deeper into the medium. However, the best image in
transillumination is obtained by subtracting the cross-polarized component Ir
from the parallel polarized component Il . In the difference Il − Ir , only ballistic
photons and weakly scattered photons, which keep their initial polarization, are
retained [4, 7].
As a second example (Figs. 5 and 6) we consider a scattering object with
Σincl = 10Σbg = 10 cm−1 and Σincl = Σbg = 0.1 cm−1 . The curves in the
(s) (s) (a) (a)
figures have the same meaning as in Figs. 3 and 4, and the time-integrated
backscattered and transmitted intensity is also shown. We obtain similar
Polarization-Based Optical Imaging 79
0.6 0.6
0.4 0.4
0 0
−2 −1 0 1 2 −2 −1 0 1 2
0.8 0.8
0.6 0.6
0.4 0.4
0 0
−2 −1 0 1 2 −2 −1 0 1 2
Distance (cm) Distance (cm)
Fig. 4. Transmission. Σbg = Σincl = 1 cm−1 , Σbg = 0.1 cm−1 , Σincl = 100 cm−1 . The
(s) (s) (a) (a)
solid lines show the total intensity I. The dashed lines show (a) the parallel-polarized
component Il , (b) the cross-polarized component Ir , (c) the polarization-difference
Il − Ir , and (d) the degree of polarization (Il − Ir )/I
Itot and Il Itot and Ir
1 1
Relative Amplitude
0.95 0.95
0.9 0.9
0.85 0.85
0.8 0.8
0.7 0.7
−2 −1 0 1 2 −2 −1 0 1 2
0.95 0.95
0.9 0.9
0.85 0.85
0.8 0.8
0.7 0.7
−2 −1 0 1 2 −2 −1 0 1 2
Distance (cm) Distance (cm)
Fig. 5. Reflection. Σincl = 10Σbg = 10 cm−1 , Σincl = Σbg = 0.1 cm−1 . The solid lines
(s) (s) (a) (a)
show the total intensity I. The dashed lines show (a) the parallel-polarized component
Il , (b) the cross-polarized component Ir , (c) the polarization-difference Il − Ir , and
(d) the degree of polarization (Il − Ir )/I
80 M. Moscoso
0.6 0.6
0.4 0.4
0 0
−2 −1 0 1 2 −2 −1 0 1 2
0.8 0.8
0.6 0.6
0.4 0.4
0 0
−2 −1 0 1 2 −2 −1 0 1 2
Distance (cm) Distance (cm)
Fig. 6. Transmission. Σincl = 10Σbg = 10 cm−1 , Σincl = Σbg = 0.1 cm−1 . The
(s) (s) (a) (a)
solid lines show the total intensity I. The dashed lines show (a) the parallel-polarized
component Il , (b) the cross-polarized component Ir , (c) the polarization-difference
Il − Ir , and (d) the degree of polarization (Il − Ir )/I
1 1
Co−polarized Intensity
(a) (b)
0.8 0.8
Total Intensity
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 5 10 15 0 5 10 15
Time Time
1 1
Cross−polarized Intensity
Polarization−Difference
(c) (d)
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 5 10 15 0 5 10 15
Time Time
Fig. 7. Time trace of the backscattered light response. Σincl = 10Σbg = 15 cm−1 ,
(s) (s)
(a) (a) −1
Σincl = Σbg
= 0.1 cm , L = 16 cm, Z = 5 cm, s = 1.6 cm. (a) Normalized total
intensity, (b) normalized co-polarized component, (c) normalized cross-polarized com-
ponent, and (d) normalized polarization-difference. The maxima in (b), (c), and (d) are
1.01, 20.36 and 1.02 times smaller than the maximum in (a). Solid curves correspond
to the signal received at the scan position (x, y) = (0, 0), just above the center of the
sphere. Dashed curves are the signal received at the scan position, (x, y) = (2.8, 0) cm,
furthest from the sphere
photons that preserve the initial polarization are present in the cross-polarized
image. The fact that the cross-polarized intensity is composed of photons that
have propagated deeper into the medium may be very useful for reconstruction
of images when one is not interested in details at and near the surface.
6 Conclusions
We have shown that simple polarization imaging techniques can be used to en-
hance the resolution of objects hidden in scattering media. Polarization imaging
is based on the fact that the polarization of the initial beam is preserved over
large optical distances. For backscattered light, we find that the cross-polarized
component shows the best resolution, while for transmitted light, images ob-
tained by subtraction of the cross-polarized component from the co-polarized
component are better. Differences between the time-dependent backscattered
signal have also been presented.
82 M. Moscoso
Acknowledgements
References
1. S. R. Arridge and J. C. Hebden, “Optical Imaging in Medicine: II. Modelling and
reconstruction,” Phys. Med. Biol. 42, pp. 841–853, 1997.
2. P. J. Michielsen K., DeRaedt H. and G. N, “Computer simulation of time-resolved
optical imaging of objects hidden in turbid media,” Phys. Reports 304, pp. 90–144,
1998.
3. S. L. Jacques, J. R. Roman, and K. Lee, “Imaging superficial tissues with polarized
light,” Laser. Surg. Med. 26(2), pp. 119–129, 2000.
4. J. M. Schmitt, A. H. Gandjbakhche, and R. F. Bonner, “Use of polarized light to
discriminate short-path photons in a multiply scattering medium,” Appl. Opt. 31,
pp. 6535–6546, 1992.
5. M. Rowe, E. Pungh, J. Tyo, and N. Engheta, “Polarization-difference imaging:
a biologically inspired technique for observation through scattering media,” Opt.
Lett. 20, pp. 608–610, 1995.
6. S. Demos and R. Alfano, “Temporal gating in highly scattering media by the degree
of optical polarization,” Opt. Lett. 21, pp. 161–163, 1996.
7. M. Moscoso, J. Keller, and G. Papanicolaou, “Depolarization and blurring of op-
tical images by biological tissues,” J. Opt. Soc. Am. A 18, No. 4, pp. 948–960,
2001.
8. S. Chandrasekhar, Radiative Transfer, Oxford University Press, Cambridge, 1960.
9. A. Kim and M. Moscoso, “Chebyshev spectral methods for radiative transfer,”
SIAM J. Sci. Comput. 23, No. 6, pp. 2075–2095, 2002.
10. B. C. Wilson and S. Jacques, “Optical reflectance and transmittance of tissues:
Principles and applications,” IEEE Quant. Electron. 26, pp. 2186–2199, 1990.
11. J. M. Schmitt and G. Kumar, “Turbulent nature of refractive-index variations in
biological tissue,” Opt. Lett. 21, pp. 1310–1312, 1996.
12. L. Ryzhik, G. Papanicolaou, and J. B. Keller, “Transport equations for elastic and
other waves in random media,” Wave Motion 24, pp. 327–370, 1996.
13. A. Kim and M. Moscoso, “Influence of the relative refractive index on depolariza-
tion of multiply scattered waves,” Physical Review E 64, p. 026612, 2001.
14. A. Kim and M. Moscoso, “Backscattering of circularly polarized pulses,” Opt.
Letters. 27, pp. 1589–1591, 2002.
15. B. Beauvoit, T. Kitai, and B. Chance, “Contribution to the mitochondrial com-
partment to the optical properties of the rat liver: a theoretical and practical
approach,” Biophys. J. 67, pp. 2501–2510, 1994.
16. J. Mourant, J. Freyer, A. Hielscher, A. Eick, D. Shen, and T. Johnson, “Mecha-
nisms of light scattering from biological cells relevant to noninvasive optical-tissue
diagnostics,” Appl. Opt. 37, pp. 3586–3593, 1998.
17. M. H. Kalos and P. A. Whitlock, Monte Carlo Methods, New York : J. Wiley Sons,
1986.
18. E. E. Lewis and W. F. Miller Jr., Computational Methods of Neutron Transport,
John Wiley and sons, New York, 1984.
Polarization-Based Optical Imaging 83
1 Introduction
A huge number of applications lead to inverse scattering problems whose goal is
to detect objects buried in a medium [2, 11, 17]. Such is the case of ultrasound
in medicine, X-ray diffraction to retrieve information about the DNA, reflection
of seismic waves in oil prospecting or crack detection in structural mechanics. A
standard procedure to find information about the obstacles consists in emitting
a certain type of radiation which interacts with the objects and the surrounding
medium and is then measured at the detector locations. The total field, con-
sisting of emitted, scattered and transmitted waves, solves a partial differential
equation (Maxwell’s equations, radiative transfer equation, equations of elastic-
ity, Helmholtz equation. . . ) with boundary conditions at the interface between
the scatterers and the medium. The original inverse problem is the following:
knowing the emitted waves and the measured patterns, find the obstacles for
which the solution of the corresponding boundary value problem agrees with the
measured values at the detector locations. There is a large literature on how
to solve it, see [8, 11] for classical results and [2, 17] for more recent advances.
86 A. Carpio and M.L. Rapún
Some of the methods proposed to solve the nonlinear inverse problem are based
on linear approximations such as the Born approximation [14], the Kirchhoff or
physical optics approximation [36] or backpropagation principles [41]. The linear
sampling method amounts to solving an equation for the contour of the unknown
domain or its representative coefficients [9, 12].
Here, we will follow the variational approach. The original formulation is
too restrictive [8, 30] and weaker formulations have been proposed in the vari-
ational setting [18, 29], leading to optimization problems with constraints given
by boundary value problems for partial differential equations. Most optimization
strategies rely on gradient principles, see for instance [33]. They deform an initial
guess for the contours of the obstacles through an iterative process in such a way
that the derivative of the functional to be minimized decreases. The techniques
differ in the type of deformation, the representation of the contours or the way
to initialize the process. The first attempts relied on classical shape optimiza-
tion using a small perturbation of the identity [27, 31, 38, 44, 52]. However, this
strategy does not allow for topological changes in the obstacle, i.e., the number
of components has to be known from the beginning. This problem was solved
by introducing deformations inspired in level-set methods [37, 50]. In this way,
contours may be created or destroyed during the iterative process. Nevertheless,
level-set based iterative methods may be rather slow unless a good initial guess
of the scatterers is available.
Topological derivative methods arise as an efficient alternative to solve inverse
scattering problems [53]. There is a growing literature on the subject, see for
instance [2, 18, 26] for scattering problems in acoustics, [21, 25] for the elasticity
case and [17, 39, 49] for problems involving electromagnetic waves. In scattering
problems, the topological derivative of the cost functional under study provides
a good first guess of the number and location of obstacles, as shown in [18, 25].
This guess may be improved using iterative schemes entirely based on topological
derivatives, as discussed in [4,21]. Hybrid level set-topological derivative methods
have been suggested in [3].
The paper is organized as follows. Section 2 discusses topological derivative
methods when the interaction between the scatterers, the medium and the in-
cident radiation is described by Helmholtz problems. We describe the forward
scattering model and reformulate the inverse problem as a constrained optimiza-
tion problem. Then, we recall the expressions for the topological derivatives of
the resulting functionals with different boundary conditions. Reconstructions of
objects obtained using topological derivatives with synthetic data are shown in
Sect. 3. Good approximations for the number, location and size of the scatterers
are found. Section 4 presents an iterative scheme which improves the description
of the shape of the obstacles in a few iterations. Section 5 gives some details
on integral methods for the numerical approximation of topological derivatives.
In Sect. 6, the analytic procedure to compute explicit expressions for the topo-
logical derivatives in Helmholtz problems is explained. Extensions to detection
of objects by means of acoustic waves in elastodynamic problems are presented
in Sect. 7. Finally, conclusions and directions for future work are discussed in
Sect. 8.
Topological Derivatives for Shape Reconstruction 87
2 Helmholtz Equations
ΓR
Γmeas
Ωi,1n
Ωi,2 n
Ωi,3n
Ωe
d
Fig. 1. Geometry of the problem: Ωi = j=1 Ωi,j and Ωe = R2 \ Ω i
88 A. Carpio and M.L. Rapún
We select the unit normal vector n pointing outside Ωe . ∂n stands for the normal
derivative at the interface and ∂r is the radial derivative. We denote by u+ and
u− the limits of u from the exterior and interior of Ωi respectively. To simplify,
the parameters λe , λi , αi are taken to be constant, real and positive. With this
assumption, we can impose the standard Sommerfeld radiation condition on the
propagation of the scattered field usc = u − uinc at infinity, which implies that
only outgoing waves are allowed.
General conditions on the parameters λe , λi , αi ensuring existence and
uniqueness of a solution u ∈ Hloc 1
(R2 ) for this problem can be found for in-
stance in [13, 32, 34, 54, 55].
Alternatively, we might reformulate (1) as a boundary value problem for the
scattered and the transmitted fields with transmission conditions utr −usc = uinc
and αi ∂n utr − ∂n usc = ∂n uinc . This is usually done for numerical purposes.
However, (1) simplifies the later computation of shape and topological derivatives
for this problem. The parameter αi is not scaled out in the second equation for
the same reason.
When the scatterer is opaque to the incident radiation (‘sound hard’ obsta-
cle), there is no transmitted wave and a Neumann boundary condition is imposed
at the interface:
⎧
⎪
⎪ ∆u + λ2e u = 0, in Ωe ,
⎪
⎪
⎨
∂n u = 0, on Γ , (2)
⎪
⎪
⎪
⎪
⎩ lim r1/2 (∂ (u − u ) − ıλ (u − u )) = 0, r = |x|.
r inc e inc
r→∞
The forward problem consists in solving (1) (resp. (2)) when the obstacles
Ωi and the incident field uinc are known. Evaluating the solution at Γmeas , we
recover the measured pattern: u|Γmeas = umeas (d). In practical experiments, the
total field u is known on the set of receptors Γmeas for different directions of
the incident wave dj . The inverse problem consists in finding the shape and
structure of the obstacle Ωi such that the solution of the forward problem (1)
(resp. (2)) equals the measured values umeas (dj ) at the receptors.
Topological Derivatives for Shape Reconstruction 89
This inverse problem is nonlinear with respect to Ωi and strongly ill-posed [8].
Given arbitrary data umeas , an associated scatterer Ωi may not exist. When it
exists, it may not depend continuously on umeas . It is well-known that the ob-
stacles are uniquely determined by the far-field pattern of the scattered wave for
all incident directions and one fixed wave number λe (see the relevant paper [30]
and the latter work [23] where the proofs in [30] for the transmission problem
were simplified). Therefore, by an analyticity argument (see [11]), if Γmeas is a
circumference, then the values of the total wave on Γmeas for all incident waves
determine uniquely the obstacles. The question of uniqueness without any a pri-
ori knowledge about the location of the obstacles for just one (or a finite number
of) incident plane waves is still an open problem.
Different strategies to regularize the inverse problem have been proposed in
the literature. We resort here to the variational approach: look for a domain Ωi
which minimizes an error in some sense. This leads to a constrained optimization
problem: minimize
1
J(Ωi ) := |u − umeas |2 dl, (3)
2 Γmeas
where u is the solution of the forward problem (1) (resp. (2)) and umeas the
measured total field on Γmeas . This functional depends on the design variable
Ωi through the boundary value problem, which acts as a constraint. When sev-
eral measurements corresponding to different directions of illumination dj are
available, the optimization problem becomes: minimize
N
1
J(Ωi ) := |uj − ujmeas |2 dl, (4)
2 j=1 Γmeas
Different notions of derivative have been introduced for shape functionals. The
topological derivative of a functional measures its sensitivity when infinitesi-
mal holes are removed from a domain. When applied to functionals associated
to inverse scattering problems such as (3), the topological derivative provides
information on the location of the scatterers: Regions where the topological
derivative takes on large negative values will be identified with places where an
object should be located.
The standard formal definition is the following. Let us consider a small ball
Bε (x) = B(x, ε), x ∈ R ⊂ R2 , and the domain Rε := R\B ε (x). The topological
derivative of J (R) is a scalar function of x ∈ R defined as
J (Rε ) − J (R)
DT (x, R) := lim . (5)
ε→0 V(ε)
90 A. Carpio and M.L. Rapún
The scalar function V(ε) is chosen in such a way that the limit (5) exists, is
finite and does not vanish. V(ε) is usually related to the measure of the ball. In
our case, V(ε) = −πε2 . In general, it will be a decreasing and negative function,
satisfying limε→0 V(ε) = 0. The value of the limit (5) depends on the partial
differential equation we use as a constraint and on the boundary conditions we
impose on the boundary of the hole. This limit also depends on the kind of hole
we are removing. It would change slightly for shapes different from a ball.
The relationship
allows to establish a link with the shape derivative (see [18]), which is useful to
obtain explicit expressions. The shape derivative of a functional J (R) along a
vector field V(z) is defined as
$
d $
DJ (R) · V := J (ϕτ (R))$$ ,
dτ τ =0
ϕτ (z) := z + τ V(z) =: zτ , z ∈ R2 .
with constant Vn < 0 and extend V to R2 in such a way that it vanishes out of
a neighborhood of ∂Bε . Then,
$
1 d $
DT (x, R) = lim J (ϕτ (R ))
ε $
$ , x ∈ R, (6)
ε→0 V (ε)|Vn | dτ
τ =0
2(1 − αi )
DT (x, R2 ) = Re ∇u(x)∇p(x) + (λ2i − λ2e ) u(x)p(x) , (7)
1 + αi
for x ∈ R2 . In all cases, u and p solve forward and adjoint problems with Ωi = ∅.
The solution u of the forward problem is the incident wave uinc (x). The adjoint
state p solves
⎧
⎪
⎨ ∆p + λ2e p = (umeas − u)δΓmeas , in R2 ,
(10)
⎪
⎩ lim r1/2 (∂r p + ıλe p) = 0,
r→∞
where δmeas is the Dirac delta function on the sampling interface Γmeas .
Theorem 2.1 holds regardless of the structure of the incident wave. It may
be a plane wave uinc (x) = exp(ı λe d · x), or a different type of source, as in
Sect. 3. Notice that the forward and adjoint solutions needed for the computation
of topological derivatives are independent of the boundary conditions at the
interface. However, they affect the final expression when calculating the limit
ε → 0.
Let us see how (7)–(9) change when we compute the topological derivative
of (3) in a domain with a hole R = R2 \ Ω, Ω being an open bounded set, not
necessarily connected. The new expression is formally identical, but the adjoint
and forward problems are solved in Ωe = R2 \ Ω and Ωi = Ω, with transmission
or Neumann boundary conditions at the interface Γ = ∂Ω. Now, the boundary
conditions not only determine the limit value but also affect the forward and
adjoint fields u and p.
Theorem 2.2. Depending on the boundary conditions at the interface, the topo-
logical derivative of the cost functional (3) in R = R2 \ Ω is given by
• General transmission problem
2(1 − αi )
DT (x, R2 \ Ω) = Re ∇u(x)∇p(x) + (λ2i − λ2e ) u(x)p(x) , (11)
1 + αi
92 A. Carpio and M.L. Rapún
where u and p solve the forward and adjoint transmission problems (1) and
(12) with Ωi = Ω and αi = 1,
• Neumann problem
" #
DT (x, R2 \ Ω) = Re 2∇u(x)∇p(x) − λ2e u(x)p(x) , (14)
where u solves the Neumann problem (2) with Ωi = Ω and p solves the adjoint
Neumann problem
⎧
⎪
⎪ ∆p + λe p = (umeas − u) δΓmeas ,
2
⎪ in Ωe ,
⎪
⎨
∂n p = 0, on Γ , (15)
⎪
⎪
⎪
⎪
⎩ lim r1/2 (∂ p + ıλ p) = 0,
r e
r→∞
for x ∈ R2 \ Ω.
The Neumann case is recovered taking the limit αi → 0, as expected [47].
Analogous formulae hold in three dimensions. The gradient term takes the
form ∇uP∇p, where P depends on the solutions of auxiliary transmission prob-
lems in a fixed ball, which change with dimension, see [24]. Inhomogeneous prob-
lems with variable parameters can be handled in a similar way, see [4].
The next sections present several tests illustrating the ability of topological
derivative based methods to locate scatterers.
where umeas (xk ) is the measured value of the total field at the k-th receptor xk
and u is the solution to the Helmholtz transmission problem associated to the
domain Ωi for an incident wave of the form (16) or (17). It is common in practice
to know the total fields for several incident waves, that is, for incident fields
corresponding to different propagation directions d1 , . . . , dN , or for different
point sources located at z1 , . . . , zN . In this case, we minimize the cost functional
(which is the discrete version of (4))
N M
1
J(Ωi ) := |uj (xk ) − ujmeas (xk )|2 , (19)
2 j=1 k=1
where uj is the solution to the forward transmission problem for the planar
incident field in the direction dj or the field generated at the source point zj ,
and ujmeas (xk ) is the measured total field at the observation point xk .
In this section, we focus on the practical situation in which one does not have
any information about the number of obstacles or its location and therefore starts
94 A. Carpio and M.L. Rapún
We are ready now to compute the topological derivative of the cost functionals.
In the 2D case, the topological derivative of (18) is
2(1 − αi )
DT (x, R2 ) = Re ∇u(x)∇p(x) + (λ2i − λ2e )u(x)p(x) ,
1 + αi
formally identical to the result in Theorem 2.1, but now the adjoint field is given
by (22). The topological derivative of the cost functional (19) is
N
2(1 − αi )
DT (x, R2 ) = Re ∇uj (x)∇pj (x) + (λ2i − λ2e )uj (x)pj (x) , (23)
j=1
1 + αi
with pj given by (22) with u and umeas replaced by uj and ujmeas , respectively.
In 3D, the factor 2(1 − αi )/(1 + αi ) should be replaced by 3(1 − αi )/(2 + αi ) in
the case of spherical holes (see [4, 24]).
If we had a priori information of the number, size and location of the ob-
stacles, we might set Ωi = ∅. In this case, u is not just the incident wave but
the solution to the Helmholtz transmission problem for Ωi . The adjoint problem
is also slightly modified, and again a transmission problem has to be solved, as
stated in Theorem 2.2. This case can be seen as the second step in an iterative
method based on topological derivatives and will be studied in Sect. 4.
Let us investigate the performance of the method when no a priori infor-
mation is available (that is, starting with Ωi = ∅) for scattering problems of
time-harmonic acoustic waves. A time-harmonic incident field has the form
Topological Derivatives for Shape Reconstruction 95
(a) (b)
3 3
−1 −4
2 2 −6
−2
−8
1 −3 1
−10
0 −4 0 −12
−14
−1 −5 −1
−16
−6 −18
−2 −2
−7 −20
−3 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
(c) (d)
3 40 3 60
30
2 2 40
20
1 1 20
10
0 0 0
0
−1 −10 −1 −20
−2 −20 −2 −40
−30
−3 −3 −60
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Fig. 2. Topological derivative with data from 24 incident directions uniformly distrib-
uted in [0,2π) (every 15◦ ) at the 14 receptors (‘×’). (a) λe = 1.25, λi = 0.25, αi = 1
(b) λe = 2.5, λi = 0.5, αi = 1 (c) λe = 5, λi = 1, αi = 1 (d) λe = 10, λi = 2, αi = 1
96 A. Carpio and M.L. Rapún
‘×’ for 24 incident planar waves with angles in [0, 2π). Incident fields were gen-
erated at different frequencies ω, increasing its values from (a) to (d). In the
four simulations we detect the presence of just one obstacle. Moreover, we guess
its location and approximate size. For some of the frequencies, the approximate
shape of the obstacle is also quite well recovered. Low and high frequencies pro-
vide different information about the obstacle. At low frequencies (see Fig. 2a,b),
the lowest values of the topological derivative are reached inside the obstacle,
located in a region of circular or elliptical shape. This allows for good recon-
structions of the location and size of the obstacle, but gives little information
about its shape. On the other hand, for high frequencies (see Fig. 2c,d), the
lowest values of the topological derivative are concentrated in a small annular
region about the boundary. Increasing the frequency, the annular region becomes
thinner. With this information, one can predict more accurately the exact shape
of the obstacle. For the same amount of data, the topological derivative is less
smooth at high frequencies (in the sense that it shows strong oscillations) than
at low frequencies. This fact makes probably more difficult to distinguish the
obstacle.
Considering a larger number of receptors and/or a larger number of incident
waves, the reconstructions for the low frequencies improve only a bit, but the
quality for the higher ones is considerably better, as can be seen in Fig. 3.
(a) (b)
3 3 −40
−10
−60
2 −20 2
−80
−30 −100
1 1
−40 −120
0 0 −140
−50
−160
−1 −60 −1 −180
−70 −200
−2 −2 −220
−80
−240
−3 −3
−90
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
(c) (d)
3 3 600
400
2 300 2 400
200 1 200
1
100
0
0 0
0
−200
−1 −100 −1
−200 −400
−2 −2
−300 −600
−3 −400 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Fig. 3. Topological derivative with data from 72 incident directions uniformly distrib-
uted in [0,2π) (every 5◦ ) at the 60 receptors (‘×’). (a) λe = 1.25, λi = 0.25, αi = 1 (b)
λe = 2.5, λi = 0.5, αi = 1 (c) λe = 5, λi = 1, αi = 1 (d) λe = 10, λi = 2, αi = 1
Topological Derivatives for Shape Reconstruction 97
(a) (b)
3 3
40
−10
2 2
20
−15
1 1
−20 0
0 0
−25 −20
−1 −1
−30 −40
−2 −2
−35 −60
−3 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
(c) (d)
3 3
−10 40
2 2
−15 20
1 1
0 −20 0 0
−1 −25 −1 −20
−2 −30 −2 −40
−3 −35 −3 −60
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Fig. 4. Topological derivative with data from 24 incident directions uniformly distrib-
uted in [0,2π) (every 15◦ ) at the 24 receptors (‘×’). (a, c) λe = 2.5, λi = 0.5, αi = 1
(b, d) λe = 5, λi = 1, αi = 1
Let us analyze the performance of the method when dealing with more com-
plex geometries. We show results for composite materials with two and three
inclusions in Fig. 4a,b and Fig. 4c,d respectively. The topological derivative is
computed for a low frequency in Fig. 4a,c and for a high one in Fig. 4b,d. The
quality of the reconstruction of the boundaries is acceptable near the observa-
tion points. The reconstruction of other parts of the boundaries is made more
difficult by the influence of each obstacle on the others. Notice that the over-
all reconstruction worsens in comparison with the case of a single obstacle. In
Fig. 4a,b we seem to detect three inclusions instead of two. The small defect is
almost undistinguishable in Fig. 4c,d. Furthermore, at low frequencies neither
shapes nor locations are well recovered.
In the previous examples we considered planar incident waves. We explore
now the choice of point sources. We repeat the experiment for the configuration
with two obstacles, keeping the same frequencies and receptors as in Fig. 4a,b,
but working now with incident fields generated at 24 uniformly distributed points
on the circumference. The location of the source points is represented in Fig. 5
with a ‘•’ symbol. The quality of the reconstruction is similar for point sources
and planar waves.
98 A. Carpio and M.L. Rapún
(a) (b)
3 −0.04 3
0.1
2 −0.06 2
−0.08 0.05
1 1
−0.1 0
0 −0.12 0
−0.05
−0.14
−1 −1
−0.16 −0.1
−2 −2
−0.18 −0.15
−3 −0.2 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Fig. 5. Topological derivative with data from 24 incident fields generated at the source
points marked with ‘•’ and measured at the 24 receptors marked with ‘×’. (a) λe =
2.5, λi = 0.5, αi = 1 (b) λe = 5, λi = 1, αi = 1
(a) (b)
3 3 0.04
0.04
2 2 0.02
0.02
1 1 0
0
0 0 −0.02
−0.02
−0.04 −0.04
−1 −1
−0.06 −0.06
−2 −2
−0.08 −0.08
−3 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
(c) (d)
3 −0.01 3 −0.01
−0.02
2 2 −0.02
−0.03
1 1 −0.03
−0.04
−0.04
0 0
−0.05
−0.05
−1 −0.06 −1
−0.06
−2 −0.07 −2
−0.07
−0.08
−3 −3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
(e) (f)
3 0.05 3 0.02
0.015
2 2 0.01
0.005
1 0 1
0
0 0 −0.005
−0.01
−1 −0.05 −1 −0.015
−0.02
−2 −2
−0.025
−3 −0.1 −3 −0.03
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Fig. 6. Topological derivative when the incident fields are generated at the source
points marked with ‘•’ and the total field is measured at the receptors marked with
‘×’. λe = 2.5, λi = 0.5, αi = 1
(a) (b)
3 3
2 2
1 1
0 0
−1 −1
−2 −2
−3 −3
2 2
2 2
0 0
0 0
−2 −2 −2 −2
(a) (b)
1 1
–5
0.8 0.8 –50
0.6 0.6
–10
0.4 0.4 –100
0.2 0.2
–15
0 0
–0.2 –150
–0.2
–20
–0.4 –0.4
–0.6 –0.6 –200
–25
–0.8 –0.8
–1 –0.1
1 0.5 1 –30 1 0.5 1 –250
0 –0.5 –1 –0.5 0 0.5 0 –0.5 –1 –0.5 0 0.5
Fig. 8. Topological derivative with data from 22 incident waves for directions
(cos θj , sin θj , 0) . The angles θj are uniformly distributed in [0, 2π). The receptors
are plotted in Fig. 7a. (a) λe = 2, λi = 1, αi = 1, (b) λe = 4, λi = 2, αi = 1
for seven angles θj uniformly distributed in [0, 2π) (the total amount of incident
waves is therefore 21, as in the previous example). The resulting topological
derivatives for the same low and high frequencies are shown in Fig. 9. In this
simple geometrical configuration with a single spherical object, both low and
high frequencies provide accurate reconstructions when choosing a suitable dis-
tribution of sampling points and incident directions.
As in the 2D case, when we consider several obstacles and more complex
shapes, the influence of one obstacle on the others as well as the appearance of
illuminated and shadow regions result in poorer reconstruction quality.
In conclusion, for simple geometries and configurations, the method produces
quite satisfactory approximations of the number, shape, size and location of the
obstacles. If we consider more complicated problems, then, the method provides
Topological Derivatives for Shape Reconstruction 101
(a) (b)
0
1 –5 1
–50
0.8 –10 0.8
0.6 0.6 –100
–15
0.4 0.4
0.2 –20 0.2 –150
0 0
–25 –200
–0.2 –0.2
–0.4 –30 –0.4
–250
–0.6 –0.6
–35
–0.8 –0.8 –300
–40 –1
–1
1 0.5 1 1 0.5 1 –350
0 –0.5 –0.5 0 0.5 –45 0 –0.5 0 0.5
–1 –1 –0.5
Fig. 9. Topological derivative with data from seven incident waves with directions
(cos θj , sin θj , 0) , seven incident waves with directions (cos θj , 0, sin θj ) and seven in-
cident waves with directions (0, cos θj , sin θj ) . The angles θj are uniformly distributed
in [0, 2π). The receptors are plotted in Fig. 7b. (a) λe = 2, λi = 1, αi = 1, (b) λe = 4,
λi = 2, αi = 1
at least a good first guess for more sophisticated iterative methods which can
circumvent the problems observed in the previous examples.
The choice of K depends on the shape of the obstacle. Our numerical experiments
suggest that, at low frequencies, the lowest values of the topological derivative are
reached inside a circular or elliptical like region and therefore we simply choose
K = 1. However, for higher frequencies, the lowest values of the topological
derivative are located in annular regions enclosing the boundaries. Depending
on the shape of the obstacle, one can decide that K = 1 is not enough and
increase its value. For the numerical experiments that we will present below, we
have taken in all cases K = 1 or K = 2 to compute initial guesses and values
Topological Derivatives for Shape Reconstruction 103
between K = 1 and K = 5 for the subsequent iterations. One could also look
for 1-periodic spline functions rj (t), which are more flexible than trigonometric
polynomials, to approximate difficult points if the original boundary Γj is smooth
but not C ∞ .
In three dimensions we can proceed in a similar way, looking for a parame-
trization in polar coordinates. Spherical harmonics play then the role of trigono-
metric polynomials.
We reproduce now the tests in Sect. 3 for the same geometries with one,
two and three scatterers. After a few iterations, the description of the obstacles
improves in a significant way.
First, we apply the iterative procedure in the configuration with a single scat-
terer. Figure 10 illustrates the performance of the method at low frequencies.
Figure 10a is the equivalent of Fig. 2b but omitting the location of the observation
points. The first guess, denoted by Ω1 , is superimposed on the representation of
(c)
(a) (b) −1
1.5
1.5 −4 1.5 −2
−2
1
1 −6 1 −3
−4
−8 0.5 −4
0.5 0.5
−10
−6 Ω
Ω1 0
2 −5
0 −12 0
−6
−14 −8 −0.5
−0.5 −0.5
−16 −7
−1 −1 −10 −1
−18 −8
−20 −1.5 −9
−1.5 −1.5 −12
−22
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
(d) (e) (f )
−1 −0.5
1.5 1.5 −1 1.5
−1
−2
1 1 1 −1.5
−2
−3 −2
0.5 0.5 0.5
−3 −2.5
Ω3 −4
Ω4 Ω5
0 0 0 −3
−5 −4 −3.5
−0.5 −0.5 −0.5 −4
−6
−5 −4.5
−1 −7 −1 −1
−5
−8 −6
−1.5 −1.5 −1.5 −5.5
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
( j) (k) (l)
1.5 −0.5 1.5 1.5
−0.5
−0.5
1 1 1
−1 −1
0.5 0.5 0.5 −1
−1.5
Ω 10 −1.5 Ω 11
Ω9
0 0 0
−2 −1.5
−2
−0.5 −0.5 −0.5
−2.5
−2.5 −2
−1 −1 −1
−3
−1.5 −3 −2.5
−1.5 −1.5
−3.5
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
Fig. 10. Topological derivative and first 11 iterations. Parameter values are λe = 2.5,
λi = 0.5 and αi = 1
104 A. Carpio and M.L. Rapún
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
1 1 2 1
0
0 0
0.5 0.5 0.5
−2
Ω3 Ω4 −2 Ω5
0 −5 0 0
−4 −4
−0.5 −0.5 −6 −0.5
−10 −6
−1 −1 −8 −1
−10 −8
−1.5 −15 −1.5 −1.5
−12
−1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5 −1.5 −1 −0.5 0 0.5 1 1.5
Fig. 11. Topological derivative and first five iterations. Parameter values are λe = 5,
λi = 1 and αi = 1
the values attained by the topological derivative when Ωi = Ω1 given in Fig. 10b.
After the first step, the magnitude of the topological derivative is almost divided
by two. In successive iterations, the order of magnitude of the updated topolog-
ical derivatives decays progressively. Their values are represented in Fig. 10c–l
for the guesses Ωi = Ωk , k = 1, . . . , 11. Ωk is the reconstruction of the obstacle
at the k-th step. Notice that the topological derivative at this step is defined
in R2 \ Ω k and Ωk appears as a solid region. The location of the obstacle is
determined in the first guess. Its approximate size and shape are satisfactorily
described after a few iterations. In 11 steps, the description of the obstacle is
almost exact.
High frequencies allow for a faster reconstruction of the obstacle in this par-
ticular geometry, as can be seen in Fig. 11. Figure 11a is the equivalent of Fig. 2c.
Notice that each approximation Ωk+1 contains the previous approximation Ωk
plus new points where the updated topological derivative falls below a threshold
−Ck , Ck > 0. A first trial value for the threshold is proposed by inspection.
We then determine Ωk+1 and update the topological derivative. In this sequence
of approximations, we observe the appearance of regions close to the boundary
where the topological derivative takes larger values (light colors). This indicates
that our current approximation may contain spurious points which do not be-
long to the obstacle. We can observe this phenomenon in Fig. 11f. In practice,
before accepting a threshold Ck , we check that the updated derivative does not
show this anomaly. Otherwise, we increase Ck .
For more complex geometries involving several scatterers, the oscillatory be-
havior of topological derivatives at high frequencies may produce patterns which
are difficult to interpret without a priori information. For this reason, we apply
the iterative procedure at low frequencies, although we know that convergence
may be slower.
Topological Derivatives for Shape Reconstruction 105
Fig. 12. Topological derivative and first eight iterations. Parameter values are λe = 2.5,
λi = 0.5 and αi = 1
The results for the geometry with two scatterers, are presented in Fig. 12.
The initial guess for the iterative scheme is computed using Fig. 4a, which seems
to indicate the presence of three obstacles, one of them misplaced. After few
iterations we find a good approximation of the real configuration. Figure 12
shows the first eight iterations. After three more iterations, the two obstacles on
the left merge and the obstacle on the right is almost recovered. As expected, the
biggest discrepancies are located in the region that is farther from the observation
points and where the influence of one obstacle on the other is stronger.
Let us now consider the geometry with three scatterers. The first guess com-
puted from Fig. 4c detects only the two largest obstacles, ignoring the small one.
For computational simplicity, if a point is included in an obstacle, it remains in-
side for subsequent iterations. In the same way, once we have created an obstacle,
we do not destroy it. It can only be merged with a neighboring one. Therefore, we
are cautious when creating a new obstacle and, although Figs. 13b,c are pointing
out the presence of the third obstacle, we have waited until the third iteration
to create it. Again, a few iterations provide a good approximation to the true
configuration.
106 A. Carpio and M.L. Rapún
Fig. 13. Topological derivative and first eight iterations. Parameter values λe = 2.5,
λi = 0.5 and αi = 1
Ωi = Ω \ B ε (x), Ωe = R2 \ Ω i , if x ∈ Ω,
Ωi = Ω ∪ Bε (x), Ωe = R2 \ Ω i , if x ∈
/ Ω.
In this way, we might implement an iterative procedure in which Ωk+1 = {x ∈
R2 | DT (x, R2 \ Ω k ) < −Ck+1 }. At each iteration, points can be added or re-
moved, see [4] for a comparison of both strategies. This would allow to make
holes inside an obstacle, for instance.
Let us describe the simple quadrature scheme for the forward and adjoint
transmission problems in Sect. 4. First of all, we rewrite the transmission problem
in terms of the scattered and the transmitted waves, which is more suitable for
numerical purposes. The function
⎧
⎪
⎨ u − uinc , in Ωe ,
v :=
⎪
⎩ u, in Ωi = ∪dj=1 Ωi,j ,
with
+i := λi /√αi .
λ
The incident field appears in the transmission conditions instead of the radiation
condition at infinity. We have switched to the standard notation in the boundary
and finite element literature where n typically stands for the exterior normal
vector pointing outside Ωi . In the literature on topological derivatives n denotes
the unit normal vector pointing inside the bounded domain Ωi . We use this
notation in Sects. 2, 6 and 7, so that the formulae agree with the ones in that
literature. We only switch to the new notation in this section.
The adjoint field p solves
⎧ M
⎪
⎪
⎪
⎪ ∆p + λ 2
p = (umeas (xk ) − u(xk )) δxk , in Ωe ,
⎪
⎪
e
⎪
⎪ k=1
⎪
⎪
⎪
⎪ ∆p + λ+2 p = 0, in Ωi,j , j = 1, . . . , d,
⎪
⎨ i
⎪
⎪ p− − p+ = 0, on Γj , j = 1, . . . , d,
⎪
⎪
⎪
⎪
⎪
⎪ α ∂ p− − ∂n p+ = 0, on Γj , j = 1, . . . , d,
⎪
⎪ i n
⎪
⎪
⎪
⎩ lim r1/2 (∂ p − ıλ p) = 0.
r e
r→∞
We decompose p as p = p1 + p2 with
⎧ N
⎪
⎪
⎨ (umeas (xk ) − u(xk )) φλe (x, xk ), in Ωe ,
p1 := k=1
⎪
⎪
⎩
0, in Ωi,j , j = 1, . . . , d,
Topological Derivatives for Shape Reconstruction 109
φλe being the fundamental solution of the Helmholtz equation (see definition
(21)). Recalling the results at the beginning of Sect. 3, the adjoint problem re-
duces to solving the following transmission problem for p2 :
⎧
⎪
⎪ 2
⎪ ∆p2 + λe p2 = 0,
⎪
in Ωe ,
⎪
⎪
⎪
⎪ +2 p2 = 0,
⎪
⎪ ∆p2 + λ in Ωi,j , j = 1, . . . , d,
⎪
⎪ i
⎨
p−2 − p2 = p1 ,
+ +
on Γj , j = 1, . . . , d, (26)
⎪
⎪
⎪
⎪
⎪
⎪ αi ∂n p−
2 − ∂ n p2 = ∂n p1 ,
+ +
⎪
⎪ on Γj , j = 1, . . . , d,
⎪
⎪
⎪
⎪
⎩ lim r1/2 (∂r p2 − ıλe p2 ) = 0.
r→∞
Notice that problems (25) and (26) are completely analogous. The only dif-
ference comes from the right hand sides. Therefore, to compute numerically the
topological derivative one just has to assemble one matrix and solve two prob-
lems with different right hand sides. We denote the right hand sides by fj and
gj . More precisely,
1 |Γj ,
fj = p+ 1 |Γj ,
gj = ∂n p+ for the adjoint problem.
In case data for several incident waves are available, a couple of problems of
the form (25) and (26) has to be solved for each of them. After discretization,
all of them will share the same matrix and only the source terms change.
Let us assume that yk : R → Γk is a smooth 1-periodic parametrization of
Γk . We define the functions
λ
Vkj (s, t) := φλ (yk (s), yj (t)),
λ
Jkj (s, t) := |yk (s)|∂nk (s) φλ (yk (s), yj (t)),
where nk (s) is the outward normal vector of Γk at the point yk (s). For the right
hand sides, we set
ϕiq,j , ϕeq,j ∈ C, q = 1, . . . , n, j = 1, . . . , d,
= fk (tp+ε ),
p = 1, . . . , n, k = 1, . . . , d, (27)
⎛ ⎞
n d n
1 i 1 + 1 1
αi ϕ + J λi (tp , tq )ϕiq,k + ⎝ ϕep,k + J λe (tp , tq )ϕeq,k ⎠
2 p,k n q=1 kk 2 n j=1 q=1 kj
1
= gk (tp ), p = 1, . . . , n, k = 1, . . . , d. (28)
n
Finally, the approximated solution to the transmission problem is computed as
⎧
⎪
⎪
d n
⎪
⎪ φλe (z, yj (tq ))ϕeq,j , z ∈ Ωe ,
⎪
⎨
ε j=1 q=1
un (z) := (29)
⎪
⎪
n
⎪
⎪ i
φλ+ i (z, yj (tq ))ϕq,j , z ∈ Ωi,j , j = 1, . . . , d.
⎪
⎩
q=1
and the densities ϕej , ϕij have to be determined. These densities are approximated
then by Dirac delta distributions on the points yj (tq ). The first set of equations
(27) is obtained when testing the transmission conditions
u− − u+ = fk , on Γk , k = 1, . . . , d,
at the observation points yk (tp+ε ), (or equivalently, we are testing the equations
with Dirac delta distributions on the points yk (tp+ε )). The second set of equa-
tions (28) is equivalent to the application of the classical Nyström method for
the second transmission conditions
Topological Derivatives for Shape Reconstruction 111
α∂n u− − ∂n u+ = gk , on Γk , k = 1, . . . , d,
which in terms of the densities are integral equations of the second kind. For
further details we refer to [15].
For any 0 = ε ∈ (−1/2, 1/2), we have a first order method. For the particular
choices ε = ±1/6, the method has quadratic convergence order, that is,
|u(z) − u±1/6
n (z)| ≤ Cz (1/n)2
This strategy was found using the ideas of qualocation methods, a variation of
collocation methods that improves the order of convergence (see [6, 16, 51]).
This method is specially well suited for our purposes because moving the
boundary requires almost no additional computational effort in recomputing the
elements in the matrix of the linear system of equations. As we have already
mentioned, this technique cannot be applied in the singular cases associated
with interior Dirichlet eigenvalues of the Laplace operator. To overcome this
problem, one can represent the solution of the transmission problem as
⎧ d
⎪
⎪
⎨ (Sjλe − ıηDjλe )ϕej , in Ωe ,
u := j=1 (32)
⎪
⎪
⎩ (S λi − ıηDλi )ϕi ,
+ +
in Ω , j = 1, . . . , d,
j j j i,j
where η > 0 is a fixed parameter, Sjλ are the single layer potentials that were
introduced in (31) and the double layer potentials Djλ are defined by
1
Djλ := |yj (t)| ∂nj (t) φλ ( · , yj (t))ϕ(t)dt.
0
112 A. Carpio and M.L. Rapún
Linear combinations of single and double layer potentials of the form S − ıηD
with η > 0 are usually called mixed or Brakhage–Werner potentials. The sys-
tem of integral equations found by imposing the transmission conditions at the
interfaces Γk is more complicated than the system obtained for the simpler rep-
resentation (30). Now, we deal with hypersingular operators. The equivalence of
this formulation to the original problem was shown in [46]. We also refer to that
work for a complete analysis of convergence of a wide class of Petrov–Galerkin
methods. In the numerical tests presented in Sect. 4, we used a fully discrete
Galerkin method with trigonometric polynomials that has superalgebraic con-
vergence order in terms of the discretization parameter, avoiding therefore the
occurrence of eigenvalues.
In 3D both integral representations (30) and (32) remain valid using the
corresponding fundamental solution in 3D given by (21). The resultant systems
of integral equations can be solved by using Petrov–Galerkin methods, as indi-
cated in [45,46]. For the implementation of fully discrete Galerkin methods with
trigonometric polynomials we refer to [40]. Discretizations of Petrov–Galerkin
methods for spline functions are studied in [45].
The shape derivative of the functional J (R) = J(Ωi ) in the direction of a smooth
vector field V(x) is defined as
$
d $
DJ(Ωi ) · V := J(φτ (Ωi ))$$ ,
dτ τ =0
Theorem 6.1. The shape derivative of the functional J defined in (3) is given by
DJ(Ωi ) · V = Re (−αi ∇u− ∇p− + λ2i u− p− + 2αi ∂n u− ∂n p− )Vn dl
Γ
− (−∇u+ ∇p+ + λ2e u+ p+ + 2∂n u+ ∂n p+ )Vn dl , (33)
Γ
where Vn = V·n, u is the solution of the forward problem and p is the continuous
solution of an adjoint problem. For particular choices of the boundary conditions
at the interface between the medium and the obstacles, (33) takes the following
simpler forms:
• General transmission problem (penetrable obstacle):
DJ(Ωi ) · V = Re αi (1 − αi )∂n u− ∂n p− + (1 − αi )∂t u− ∂t p− Vn dl
Γ
+ (λ2i − λ2e ) u p Vn dl . (34)
Γ
where u and p solve the forward and adjoint transmission problems (1) and
(12) with αi = 1.
• Neumann problem (sound hard obstacle):
DJ(Ωi ) · V = Re (∇u∇p − λ2e up)Vn dl , (36)
Γ
by parts the resulting expression involving adjoint solutions, we find the desired
formula for the shape derivative. The proof is organized in four steps.
Step 1: Variational formulation of the forward problem. First, we replace
the exterior problem (1) by an equivalent boundary value problem posed in
a bounded domain. Let us introduce a circle ΓR which encloses the obstacles
as in Fig. 1. The Dirichlet-to-Neumann (also called Steklov–Poincaré) operator
associates to any Dirichlet data on ΓR the normal derivative of the solution of
the exterior Dirichlet problem:
where w ∈ Hloc
1
(R2 \ B R ), BR := B(0, R), is the unique solution of
⎧
⎪
⎪ ∆w + λ2e w = 0, in R2 \ B R ,
⎪
⎪
⎨
w = f, on ΓR ,
⎪
⎪
⎪
⎪
⎩ lim r1/2 (∂ w − ıλ w) = 0.
r e
r→∞
1
Hloc (R2 \B R ) denotes the usual Sobolev space and H 1/2 (ΓR ) and H −1/2 (ΓR ) are
the standard trace spaces. A detailed representation of the Dirichlet-to-Neumann
map in terms of Hankel functions is given in [19, 28]. Instead of the exterior
transmission problem (1) one can study an equivalent boundary value problem
in BR with a non-reflecting boundary condition on ΓR :
Topological Derivatives for Shape Reconstruction 115
⎧
⎪
⎪ 2
in Ωe := BR \ Ω i ,
⎪ ∆u + λe u = 0,
⎪
⎪
⎪
⎪
⎪
⎪
⎪ αi ∆u + λ2i u = 0, in Ωi ,
⎪
⎪
⎨
u− − u+ = 0, on Γ , (37)
⎪
⎪
⎪
⎪
⎪
⎪ αi ∂n u− − ∂n u+ = 0,
⎪
⎪ on Γ ,
⎪
⎪
⎪
⎪
⎩ ∂n (u − uinc ) = L(u − uinc ), on ΓR .
where
b(Ωi ; u, v) := (∇u∇v − λ2e uv)dz + (αi ∇u∇v − λ2i uv)dz
Ωe Ωi
− Lu v dl, ∀u, v ∈ H 1 (BR ),
ΓR
(v) := (∂n uinc − Luinc ) v dl, ∀v ∈ H 1 (BR ).
ΓR
with
b(Ωi,τ ; u, v) := (∇zτ u∇zτ v − λ2e uv)dzτ + (αi ∇zτ u∇zτ v − λ2i uv)dzτ
Ωe,τ Ωi,τ
− Lu vdl, ∀u, v ∈ H 1 (BR ).
ΓR
d
where u̇ = dτ uτ |τ =0 . To avoid computing u̇ we introduce an adjoint problem.
Let us define a family of modified functionals
thanks to (41) and (42). This is true for any p ∈ H 1 (ΩR ). If p solves
⎧
⎪
⎨ p ∈ H (BR ),
1
(44)
⎪
⎩ b(Ωi ; v, p) = (umeas − u) v dl, ∀v ∈ H 1 (BR ),
Γmeas
the derivative u̇ is not needed and the shape derivative DJ(Ω) · V is given by
the first term in (43). We select the solution p of (44) as an adjoint state. The
adjoint problem (44) is equivalent to
⎧
⎪
⎨ p ∈ H (BR ),
1
(45)
⎪
⎩ b(Ωi ; p, v) = (umeas − u) v dl, ∀v ∈ H 1 (BR ),
Γmeas
which has the same structure as the direct problem (38) and p is then a solution
of (12). A property of the Dirichlet-to-Neumann operator is essential in this
argument: its adjoint operator satisfies L∗ (u) = L(u).
Step 4: Shape derivative. Let us take p to be the adjoint state and compute
the first term in (43). The change of variables from the deformed to the original
variables is governed by the identities
Topological Derivatives for Shape Reconstruction 117
and
$ $ $
d $ d $ d $
Fτ $$ = ∇V, (∇zτ u)$$ = −∇V ∇u, dzτ $$ = ∇ · Vdz,
dτ τ =0 dτ τ =0 dτ τ =0
and
− (∇V + ∇V )∇u · ∇p dz =
Ωi
∂ 2 u ∂p ∂ 2 p ∂u ∂ 2 u ∂p ∂ 2 p ∂u
2 v + v dz + v + v dz
Ωi ∂xj ∂x ∂x2j ∂x Ωi ∂xj ∂x ∂xj ∂xj ∂x ∂xj
∂u− ∂p− ∂p− ∂u−
+ v nj + v nj dz, (48)
Γ ∂xj ∂x ∂xj ∂x
because u and p satisfy Helmholtz equations in Ωi . The sum of the integrals over
Γ is
118 A. Carpio and M.L. Rapún
−αi (∇u− · ∇p− )Vn dl + λ2i u− p− Vn dl + 2αi (∇u− · n)(∇p− · n)Vn dl.
Γ Γ Γ
using V = Vn n.
We reproduce the same computations for Ωe . All the integrals over ΓR are
identically zero, since V = 0 on ΓR . Recall that n is now the outward normal
vector. Integrating by parts we obtain
(∇u · ∇p − λ2e up) ∇ · V dz − (∇V + ∇V )∇u · ∇p dz
Ωe Ωe
= (∆u + λ2e u)∇p · V dz + (∆p + λ2e p)∇u · V dz +
Ωe Ωe
(∇u ∇p )Vn dl −
+ +
λ2e u p Vn dl − 2
+ +
(∇u+ · n)(∇p+ · n)Vn dl. (49)
Γ Γ Γ
Due to these identities and the fact that V = 0 on Γmeas , the two integrals over
Ωe vanish.
Adding all the integrals on Γ generated integrating by parts on Ωi and Ωe ,
we finally find (33). 2
Theorem 2.1.- Transmission problem without holes. Let us first calculate the
topological derivative of the cost functional (3) for the transmission problem
when R = R2 . Knowing the formula of the shape derivative for the transmission
problem, we use the identities (6) and (34) to obtain
−1
DT (x, R2 ) = lim Re αi (1 − αi )∂n u− −
ε ∂n pε (50)
ε→0 V (ε) Γε
− −
+ (1 − αi )∂t uε ∂t pε dl + (λ2i − λ2e )uε pε dl .
Γε
Recall that Vn is constant and negative. Here, uε and pε solve the forward and
adjoint problems when Ωi = Bε (x) and Γ = Γε = ∂Bε (x). A value for this limit
is computed performing an asymptotic expansion of these solutions and their
gradients at Γε as in [4].
The asymptotic behavior of uε and pε is obtained expressing them as correc-
tions of u and p
Let us expand now vε (ξ) in powers of ε : vε (ξ) = v (1) (ξ) + εv (2) (ξ) + O(ε2 ).
The leading terms of the expansion solve:
⎧
⎪
⎪ ∆ξ v (1) = 0, in R2 \ B and B,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨ v (1)− − v (1)+ = u(x), on Γ ,
(52)
⎪
⎪
⎪ αi n(ξ) · ∇ξ v
⎪
(1)−
− n(ξ) · ∇ξ v (1)+
= 0, on Γ ,
⎪
⎪
⎪
⎪
⎪
⎩ lim r1/2 ∂r v (1) = 0,
r→∞
and ⎧
⎪
⎪ ∆ξ v (2) = 0, in R2 \ B and B,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨ v (2)− − v (2)+ = ξ · ∇u(x), on Γ ,
(53)
⎪
⎪
⎪
⎪ α n(ξ) · ∇ξ v (2)− − n(ξ) · ∇ξ v (2)+ = n(ξ) · ∇u(x), on Γ ,
⎪ i
⎪
⎪
⎪
⎪
⎩
limr→∞ r1/2 ∂r v (2) − ıλe v (1) = 0.
By inspection, v (1) (ξ) = u(x)χB (ξ). The second term, v (2) (ξ), can be found
working in polar coordinates:
2 1 − αi ξ
v (ξ) = ∇u(x) ·
(2)
ξ χB (ξ) + χ 2 (ξ) = ∇u(x) · g(ξ).
1 + αi 1 + αi |ξ|2 R \B
Thus,
∂u−
ε ∂v (2)− ∂v (2)− 2 ∂u
(z) = ε (ξ) + O(ε) = (ξ) + O(ε) = (x) + O(ε),
∂zj ∂zj ∂ξj 1 + αi ∂zj
and a similar identity for pε . Therefore,
2 2
∇u−
ε (z) → ∇u(x), ∇p−
ε (z) → ∇p(x), as ε → 0.
1 + αi 1 + αi
Let us take limits in the three integrals appearing in (50). For the integral
of the normal components, we write the unit normal as a function of the angle
(n1 , n2 ) = (cos θ, sin θ). Then,
2π
4 ∂uε ∂pε
(∇u−
ε · n)(∇p −
ε · n)dl z ≈ ε (x) (x) nj n dθ
Γε (1 + αi )2 ∂zj ∂z 0
4π
= ε ∇u(x)∇p(x),
(1 + αi )2
( 2π
plus a remainder of higher order in ε, where we have used that 0 n1 n2 dθ =
( 2π
0 and 0 n2j dθ = π for j = 1, 2. To compute the integral of the tangential
components, we write the tangent vector as (t1 , t2 ) = (− sin θ, cos θ). Exactly
the same value is found. Finally, the third integral is
uε pε dlz ≈ 2επ u(x)p(x).
Γε
Taking into account that V (ε) = −2πε, (7) follows. Formula (8) is a particular
case with αi = 1.
Theorem 2.1.- Neumann problem without holes. The analytical expression of
the topological derivative for Neumann problems (9) follows by combining (6)
and (36). Then, we perform an asymptotic expansion of uε and pε , which now
solve Neumann problems. We find that
ξ
uε (z) = u(z) + ε∇u(x) · + O(ε2 ).
|ξ|2
Theorem 2.2.- Transmission problem with holes. Let us calculate now the
topological derivative of the cost functional (3) in a domain with a hole R =
R2 \ Ω, Ω being an open bounded set, not necessarily connected. Formula (11)
follows by slightly modifying the procedure we have used to compute (7) in R2
for the transmission problem.
Now, uε (z) = u(z)χR2 \B (z) + vε (z), where u solves (1) with Ωi = Ω. Chang-
ing variables, vε (ξ) satisfies
⎧
⎪
⎪ ∆ξ vε + ε2 λ2e vε = 0, in R2 \ (B ∪ Ω ε ),
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ 2 2
in B ∪ Ωε ,
⎪ αi ∆ξ vε + ε λi vε = 0,
⎪
⎪
⎪
⎪
⎪
⎪
⎪ vε− − vε+ = u(z) = u(x) + ε ξ · ∇u(x) + O(ε2 ), on Γ ,
⎪
⎪
⎪
⎨
αi n(ξ) · ∇ξ vε− − n(ξ) · ∇ξ vε+ = εn(ξ) · ∇u(x) + O(ε2 ), on Γ ,
⎪
⎪
⎪
⎪
⎪
⎪
⎪ vε− − vε+ = 0,
⎪ on ∂Ωε ,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ αi n(ξ) · ∇ξ vε− − n(ξ) · ∇ξ vε+ = 0, on ∂Ωε ,
⎪
⎪
⎪
⎪
⎪
⎩ lim r1/2 (∂ v − ıελ v ) = 0,
r ε e ε r = |ξ|,
r→∞
(54)
with Ωε := (Ω − x)/ε.
Expanding vε (ξ) in powers of ε, the leading terms v (1) (ξ), v (2) (ξ) solve again
(52) and (53), respectively, and (11) follows. We just have to check that the
presence of Ωε does not provide corrections to the orders zero and one. Let us
consider the boundary value problems:
⎧
⎪
⎪ ∆ξ v (1) = 0, in R2 \ (B ∪ Ω ε ) and B ∪ Ωε ,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ (1)−
− v (1)+ = u(x),
⎪v
⎪
on Γ ,
⎪
⎪
⎪
⎪
⎪
⎨ αi n(ξ) · ∇ξ v (1)− − n(ξ) · ∇ξ v (1)+ = 0, on Γ ,
(55)
⎪
⎪
⎪
⎪ v (1)−
− v (1)+
= 0, on ∂Ω ,
⎪
⎪
ε
⎪
⎪
⎪
⎪
⎪
⎪ αi n(ξ) · ∇ξ v
(1)−
− n(ξ) · ∇ξ v (1)+ = 0, on ∂Ωε ,
⎪
⎪
⎪
⎪
⎪
⎩ lim r1/2 ∂ v (1) = 0,
r
r→∞
and
122 A. Carpio and M.L. Rapún
⎧
⎪
⎪ ∆ξ v (2) = 0, in R2 \ (B ∪ Ω ε ) and B ∪ Ωε ,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ v (2)− − v (2)+ = ξ · ∇u(x), on Γ ,
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎨ αi n(ξ) · ∇ξ v (2)− − n(ξ) · ∇ξ v (2)+ = n(ξ) · ∇u(x), on Γ ,
(56)
⎪ (2)−
⎪
⎪
⎪ v − v (2)+
= 0, on ∂Ω ,
⎪
⎪
ε
⎪
⎪
⎪
⎪
⎪
⎪ αi n(ξ) · ∇ξ vε
(2)−
− n(ξ) · ∇ξ v (2)+ = 0, on ∂Ωε ,
⎪
⎪
⎪
⎪
⎪
⎩ lim (2)
r→∞ r
1/2
∂r v − ıλe v (1) = 0.
solves (56) with an error of order ε2 . Thus, the correction coming from Ωε
appears at orders higher than ε.
A similar argument works for the Neumann problem.
The ideas and techniques developed in the previous sections for scattering prob-
lems involving waves governed by Helmholtz equations can be extended to more
general situations. In this section, we consider the problem of detecting solids
buried in an elastic medium by means of acoustic waves. We only describe the
procedure to compute explicit formulae for the topological derivatives in this new
setting. Once these formulae are known, the numerical approximation procedures
described in Sects. 3–5 apply using the fundamental solutions and integral oper-
ators of elasticity theory.
where cjαmβ are the stiffness tensor components, ρ is the mass density and z
and t are the 2D position vector and time, respectively. Subscripts range from 1
to 2. We adopt the convention that a repeated index is summed over its range.
Regardless of particular material symmetries, the elastic moduli satisfy the
following symmetry restrictions:
for some constant C > 0 and every complex ξjα such that ξjα = ξαj . The elastic
constants cjαmβ for a cubic crystal are:
cjαmβ = c12 δjα δmβ + c44 (δjm δαβ + δjβ δαm ) − Hδjαmβ (57)
where H is the anisotropy factor H = 2c44 + c12 − c11 . δjm stands for the
Kronecker delta. In the isotropic case, c12 = λ, c44 = µ and c11 = λ + 2µ.
We are interested in time harmonic solutions, in which the elastic displace-
ment field can be written in the form Uj (z, t) = Re [uj (z)e−ıωt ] for a given
frequency ω > 0. Therefore, the components uj , j = 1, 2 solve
∂ 2 um
ρω 2 uj + cjαmβ = 0. (58)
∂zα ∂zβ
In the scattering process, a time harmonic vector plane wave of unit ampli-
tude uinc illuminates an object with section Ωi embedded in an infinite elastic
medium. This process generates two additional time harmonic fields: the scat-
tered vector field usc , defined outside the obstacle and propagating outwards,
together with the transmitted vector field utr , defined inside the obstacle and
trapped in it.
Dropping the time harmonic dependence, which is implicit in (58), the inci-
dent waves have the form uinc (z) = u0 eık·z , with z = (z1 , z2 ) and k = (k1 , k2 ) =
kd, where k is the wave number and the unitary vector d is the direction of
propagation. The wave vector k, the frequency ω and u0 are related by:
Note that the frequency ω depends on the wave vector k through the dispersion
relation (60). Substituting each of its two roots ω 2 = ωj2 (k) in system (59), we
find the directions of displacement u0 in these waves (directions of polarization).
124 A. Carpio and M.L. Rapún
satisfies:
⎧
⎪ ∂ 2 um
⎪
⎪ ρi 2
ω u + ci
= 0, j = 1, 2, in Ωi ,
⎪
⎪
j jαmβ
∂zα ∂zβ
⎪
⎪
⎪
⎪
⎪
⎪ ∂ 2 um
⎪
⎨ ρe ω 2 uj + cejαmβ = 0, j = 1, 2, in Ωe ,
∂zα ∂zβ
(61)
⎪
⎪
⎪ u− − u+ = 0,
⎪
⎪
⎪ on Γ := ∂Ωi ,
⎪
⎪
⎪
⎪
⎪ i
⎪ ∂u− ∂u+
⎩ cjαmβ m nα − cejαmβ m nα = 0, j = 1, 2, on Γ ,
∂zβ ∂zβ
lim r1/2 (∂r (uc − ucinc ) − ıkc (uc − ucinc )) = 0, c = p, s r = |x|. (62)
r→∞
Topological Derivatives for Shape Reconstruction 125
Solutions satisfying this condition are called radiating solutions of the isotropic
2D Navier equations, see the reviews [22, 35] and references therein. Radiation
conditions for anisotropic materials are discussed in [22].
For rigid obstacles, there is no transmitted field. Then, the transmission
problem is replaced by a Neumann problem:
⎧
⎪ ∂ 2 um
⎪
⎪
e 2 e
⎨ ρ ω uj + cjαmβ ∂zα ∂zβ = 0, j = 1, 2, in Ωe ,
(63)
⎪
⎪ ∂um
⎪ e
⎩ cjαmβ nα = 0, j = 1, 2, on Γ ,
∂zβ
with a ‘sound hard’ boundary condition at the obstacle boundary plus the radi-
ation condition on u − uinc at infinity. For simplicity, in this section we will work
only with the Neumann problem and we will adopt the simpler notation ρ ≡ ρe
and cjαmβ ≡ cejαmβ . Similar, but more involved computations can be carried out
for the transmission problem (see [5]).
The forward problem consists in computing the solution u of (63) at the
measurement curve Γmeas knowing the scatterers Ωi and the incident wave uinc .
Existence and uniqueness results for problems (61) and (63) can be found in
[1, 7, 20, 25, 35, 42, 43].
As before, the inverse problem consists in finding the shape and structure
of the obstacle Ωi such that the solution of the forward Neumann problem (63)
equals the measured values umeas at the receptors, knowing the incident field.
A less restrictive formulation is given by the following constrained optimization
problem: minimize
1
J(Ωi ) := |u − umeas |2 (64)
2 Γmeas
where u is the solution of the forward Neumann problem (63) and umeas the
total field measured on Γmeas .
The computation of the topological derivative of the shape functional (64) follows
the same steps as in Sect. 6. Let us first compute the shape derivative of (64).
Then, we will use its expression to find an explicit formula for the topological
derivative.
Theorem 7.1. Keeping the notations of the previous sections, the shape deriv-
ative of the functional J defined in (64) is given by
∂uj ∂pm
DJ(Ω) · V = Re cjαmβ − ρω 2 up Vn dl , (65)
Γ ∂zα ∂zβ
where
∂uj ∂v m
b(Ωi ; u, v) := (cjαmβ − ρω 2 uv) dz − Lu v dl,
Ωe ∂zα ∂zβ ΓR
∂uinc,m
(v) := (cjαmβ nα v j − Luinc v) dl, ∀v ∈ (H 1 (BR ))2 .
ΓR ∂zβ
Step 2: Transformed problems. Let uτ be the solution of the variational for-
mulations in the transformed domains:
⎧
⎪
⎨ uτ ∈ (H 1 (BR ))2 ,
(69)
⎪
⎩ b(Ω ; u , v) = (v),
i,τ τ ∀v ∈ (H (BR )) ,
1 2
with
∂uj ∂v m
b(Ωi,τ ; u, v) := (cjαmβ − ρω 2 uv) dzτ
Ωe,τ ∂zτ,α ∂zτ,β
− Lu v dlτ , ∀u, v ∈ (H 1 (BR ))2 .
ΓR
the derivative u̇ is not needed and the shape derivative DJ(Ω) · V is given by
the first term in (73). Problem (74) is the adjoint problem of (68).
The definition of Steklov–Poincaré operator implies that L∗ (p) = L(p).
Thus, the adjoint problem is equivalent to:
⎧
⎪
⎪
⎨ p ∈ (H (BR )) ,
1 2
(75)
⎪
⎪ (umeas − u) v, ∀v ∈ (H 1 (BR ))2 .
⎩ b(Ωi ; p, v) =
Γmeas
Integrating by parts, taking into account that n is the outward normal vector
to Ωe and that V vanishes on ΓR
$
d $ ∂uj ∂pm
$
b(Ωi,τ ; u, p)$ = (cjαmβ − ρω 2 up) Vn dl
dτ τ =0 Γ ∂zα ∂zβ
∂ 2 uj ∂pm ∂uj ∂ 2 pm ∂u ∂p
+ cjαmβ ( + ) − ρω 2 ( p+u ) Vγ dz
Ωe ∂z α ∂z γ ∂z β ∂z α ∂z β ∂z γ ∂z γ ∂z γ
∂ 2 uj ∂pm ∂uj ∂ 2 pm ∂ 2 uj ∂pm ∂uj ∂ 2 pm
− cjαmβ ( + + + ) Vγ dz
Ωe ∂zα ∂zγ ∂zβ ∂zγ ∂zβ ∂zα ∂zα ∂zβ ∂zγ ∂zα ∂zβ ∂zγ
∂u− −
j ∂pm ∂u− −
j ∂pm
+ cjαmβ (nα + nβ )Vn nγ dl. (76)
Γ ∂zγ ∂zβ ∂zα ∂zγ
The sum of all the integrals over Ωe that appear in (76) is
∂ 2 uj ∂p ∂ 2 pm ∂uj
− (cjαmβ +ρω 2 um ) m + (cjαmβ +ρω 2 pj ) Vγ dz = 0
Ωe ∂z α ∂z β ∂z γ ∂z β ∂z α ∂zγ
Topological Derivatives for Shape Reconstruction 129
because
∂ 2 um ∂ 2 pm
ρω 2 uj + cjαmβ = 0, ρω 2 pj + cjαmβ = (umeas − u)j δΓmeas , in Ωe
∂zα ∂zβ ∂zα ∂zβ
and cjαmβ = cmβjα . The field V = 0 on Γmeas , thus all the integrals over Γmeas
are identically zero.
Adding the integrals on Γ we find
∂uj ∂pm
cjαmβ − ρω 2 up Vn dl
Γ ∂zα ∂zβ
∂uj ∂pm ∂uj ∂pm
+ cjαmβ nα + nβ Vn nγ dl.
Γ ∂zγ ∂zβ ∂zα ∂zγ
Using the homogeneous Neumann condition, the second integral vanishes and
(65) follows. 2
pq pq
where w(ξ) = cpqrs ∂u
∂zs (x)θ (ξ), and θ
r
solve the problems
⎧
⎪
⎪ ∂ 2 θm
pq
⎪
⎨ cjαmβ ∂ξ ∂ξ = 0, j = 1, 2, in R2 \ B(0; 1),
α β
(81)
⎪
⎪ ∂θpq
⎪ pq
⎩ −cjαmβ m nα = Mjα nα , j = 1, 2, on Γ ,
∂ξβ
130 A. Carpio and M.L. Rapún
and write wε = w(1) + εw(2) + O(ε2 ). The leading term, w(1) is the radiating
solution of ⎧ (1)
⎪
⎪ ∂ 2 wm
⎪ cjαmβ
⎪ = 0, in R2 \ B,
⎨ ∂ξα ∂ξβ
(84)
⎪
⎪ (1)
⎪
⎪
∂wm
⎩ −cjαmβ nα = 0, on Γ .
∂ξβ
The solution of this problem is w(1) (ξ) = 0. The second term, w(2) , is the
radiating solution of
⎧
⎪
⎪
(2)
∂ 2 wm
⎪
⎪ in R2 \ B,
⎨ jαmβ ∂ξα ∂ξβ = 0,
c
(85)
⎪
⎪ (2)
⎪
⎪ ∂w m ∂u m
⎩ −cjαmβ nα = cjαmβ (x)nα , on Γ .
∂ξβ ∂zβ
Topological Derivatives for Shape Reconstruction 131
Notice that, since the interface Γ is the unit circle, n = ξ/|ξ| = ξ. The function
w(2) can be written in terms of the elementary solutions of (81) as w(2) (ξ) =
pq
cpqrs ∂u
∂zs (x)θ (ξ). Our expansion yields:
r
8 Conclusions
Topological derivative methods are a powerful tool to devise efficient numerical
schemes for solving inverse scattering problems. In problems where the incident
radiation on the obstacles is governed by Helmholtz or elasticity equations, we
have computed expressions for the topological derivatives of appropriate con-
strained optimization regularizations. Topological derivative based numerical
strategies to approximate the scatterers have been discussed. We have performed
a number of tests in 2D and 3D illustrating the advantages and disadvantages of
the method. In general, fairly good approximations of the number, size, location
and shape of the scatterers are obtained at a low computational cost.
Many real situations require not only information about the scatterers, but
also about constitutive parameters that we assume here to be known. An ex-
tension of topological derivative methods to provide information on both the
obstacles and their parameters is proposed in [4].
Acknowledgements
This research has been supported by Grants MAT2005-05730-C02-02 and MTM-
2004-01905 of the Spanish MEC, BSCH/UCM PR27/05-13939 and CM-910143.
References
1. Anagnostopoulos KA and Charalambopoulos A 2006 The linear sampling method
for the transmission problem in two-dimensional anisotropic elasticity Inverse prob-
lems 22 553–577
2. Bonnet M and Constantinescu A 2005 Inverse problems in elasticity Inverse prob-
lems 21 R1-R50
3. Burger M, Hackl B and Ring W 2004 Incorporating topological derivatives into
level set methods J. Comp. Phys. 194 344–362
4. Carpio A and Rapun ML, Solving inhomogeneous inverse problems by topological
derivative methods, submitted, 2007
5. Carpio A and Rapun ML, Shape reconstruction in anisotropic elastic media. In
preparation
6. Chandler GA and Sloan IH 1990. Spline qualocation methods for boundary integral
equations. Numer. Math. 58, 537–567
132 A. Carpio and M.L. Rapún
32. Kleinman RE and Martin P 1988 On single integral equations for the transmission
problem of acoustics SIAM J. Appl. Math 48 307–325
33. Kleinman RE and van der Berg PM 1992 A modified gradient method for two
dimensional problems in tomography J. Comput. Appl. Math. 42 17–35
34. Kress R and Roach GF 1978 Transmission problems for the Helmholtz equation
J. Math. Phys. 19 1433–1437
35. Kupradze VD, Gegelia TG, Basheleuishvili MO and Burchauladze TV, Three di-
mensional problems of the mathematical theory of elasticity and thermoelasticity,
North-Holland Ser. Appl. Math. Mech. 25, North-Holland, Amsterdam, 1979.
36. Liseno A and Pierri R 2004 Imaging of voids by means of a physical optics based
shape reconstruction algorithm J. Opt. Soc. Am. A 21 968–974
37. Litman A, Lesselier D and Santosa F 1998 Reconstruction of a two-dimensional
binary obstacle by controlled evolution of a level set Inverse problems 14 685–706
38. Masmoudi M 1987 Outils pour la conception optimale des formes Thèse d’Etat en
Sciences Mathématiques, Université de Nice
39. Masmoudi M, Pommier J and Samet B 2005 The topological asymptotic expansion
for the Maxwell equations and some applications Inverse Problems 21 547–564
40. Meddahi S and Sayas FJ 2005 Analysis of a new BEM–FEM coupling for two
dimensional fluid–solid iteraction Num. Methods Part. Diff. Eq. 21 1017–1154
41. Natterer F and Wubbeling F 1995 A propagation backpropagation method for
ultrasound tomography Inverse problems 11 1225-1232
42. Natroshvili D 1995 Mixed interface problems for anisotropic elastic bodies Georgian
Math. J. 2 631-652
43. Natroshvili D 1996 Two-dimensional steady state oscillation problems of
anisotropic elasticity Georgian Math. J. 3 239-262
44. Potthast R 1996 Fréchet differentiability of the solution to the acoustic Neumann
scattering problem with respect to the domain J. Inverse Ill-Posed Problems 4
67–84
45. Rapún ML and Sayas FJ 2006 Boundary integral approximation of a heat diffusion
problem in time–harmonic regime Numer. Algorithms 41 127–160
46. Rapún ML and Sayas FJ 2006 Indirect methods with Brakhage–Werner potentials
for Helmholtz transmission problems. In Numerical Mathematics and advanced
applications. ENUMATH 2005. Springer 1146–1154
47. Rapún ML and Sayas FJ 2007 Exterior Dirichlet and Neumann problems for the
Hemholtz equation as limits of transmission problems. Submitted.
48. Rapún ML and Sayas FJ 2007 Boundary element simulation of thermal waves.
Arch. Comput. Methods. Engrg 14 3–46
49. Samet B, Amstutz S and Masmoudi M 2003 The topological asymptotic for the
Helmholtz equation SIAM J. Control Optimim 42 1523–1544
50. Santosa F 1996 A level set approach for inverse problems involving obstacles
ESAIM Control, Optim. Calculus Variations 1 17–33
51. Sloan IH 2000. Qualocation. J. Comput. Appl. Math. 125, 461–478
52. Sokolowski J and Zolésio JP 1992 Introduction to shape optimization. Shape sen-
sitivity analysis (Heidelberg: Springer)
53. Sokowloski J and Zochowski A 1999 On the topological derivative in shape opti-
mization SIAM J. Control Optim. 37 1251–1272
54. Torres RH and Welland GV 1993 The Helmholtz equation and transmission prob-
lems with Lipschitz interfaces Indiana Univ. Math. J. 42 1457–1485
55. von Petersdorff T 1989 Boundary integral equations for mixed Dirichlet, Neumann
and transmission problems. Math. Methods Appl. Sci. 11 185–213
Time-Reversal and the Adjoint Imaging Method
with an Application in Telecommunication
Oliver Dorn
1 Introduction
Time-reversal techniques have attracted great attention recently due to the large
variety of interesting potential applications. The basic idea of time-reversal (often
also referred to as ‘phase-conjugation’ in the frequency domain) can be roughly
described as follows. A localized source emits a short pulse of acoustic, electro-
magnetic or elastic energy which propagates through a richly scattering envi-
ronment. A receiver array records the arriving waves, typically as a long and
complicated signal due to the complexity of the environment, and stores the
time-series of measured signals in memory. After a short while, the receiver ar-
ray sends the recorded signals time-reversed (i.e. first in – last out) back into
the same medium. Due to the time-reversibility of the wave fields, the emitted
energy backpropagates through the same environment, practically retracing the
paths of the original signal, and refocuses, again as a short pulse, on the loca-
tion where the source emitted the original pulse. This, certainly, is a slightly
oversimplified description of the rather complex physics which is involved in the
136 O. Dorn
real time-reversal experiment. In practice, the quality of the refocused pulse de-
pends on many parameters, as for example the randomness of the background
medium, the size and location of the receiver array, temporal fluctuations of
the environment, etc. Surprisingly, the quality of the refocused signal increases
with increasing complexity of the background environment. This was observed
experimentally by M. Fink and his group at the Laboratoire Ondes et Acous-
tique at Université Paris VII in Paris by a series of laboratory experiments (see
for example [17, 20, 21]), and by W. A. Kuperman and co-workers at the Scripps
Institution of Oceanography at University of California, San Diego, by a series of
experiments performed between 1996 and 2000 in a shallow ocean environment
(see for example [9, 18, 26]).
The list of possible applications of this time-reversal technique is long. In an
iterated fashion, resembling the power method for finding maximal eigenvalues
of a square matrix, the time-reversal method can be applied for focusing energy
created by an ultrasound transducer array on strongly scattering objects in the
region of interest. This can be used for example in lithotripsy for localizing and
destroying gall-stones in an automatic way, or more generally in the application
of medical imaging problems. Detection of cracks in the aeronautic industry, or
of submarines in the ocean are other examples. See for example [39, 40], and for
related work [7,12]. The general idea of time-reversibility, and its use in imaging
and detection, is certainly not that new. Looking into the literature for example
of seismic imaging, the application of this basic idea can be found in a classical
and very successful imaging strategy for detecting scattering interfaces in the
Earth, the so-called ‘migration’ technique [4,8]. However, the systematic use and
investigation of the time-reversal phenomenon and its experimental realizations
started more recently, and has been carried out during the last 10–15 years or
so by different research groups. See for example [17, 18, 20, 21, 26, 31, 46, 49] for
experimental demonstrations, and [1–3,5,6,9,15,16,23,28,30,38,39,43–45,47,50]
for theoretical and numerical approaches.
One very young and promising application of time-reversal is communication.
In these lecture notes we will mainly concentrate on that application, although
the general results should carry over also to other applications as mentioned
above.
The text is organized as follows. In Sect. 2 we give a very short introduction
into time-reversal in the ocean, with a special emphasis on underwater sound
communication. Wireless communication in a MIMO setup, our second main
application in these notes, is briefly presented in Sect. 3. In Sect. 4 we discuss
symmetric hyperbolic systems in the form needed here, and examples of such
systems are given in Sect. 5. In Sect. 6, the basic spaces and operators necessary
for our mathematical treatment are introduced. The inverse problem in com-
munication, which we are focusing on in these notes, is then defined in Sect. 7.
In Sect. 8, we derive the basic iterative scheme for solving this inverse problem.
Section 9 gives practical expressions for calculating the adjoint communication
operator, which plays a key role in the iterative time-reversal schemes. In Sect. 10
the acoustic time-reversal mirror is defined, which will provide the link between
Time-Reversal and the Adjoint Imaging Method in Communication 137
medium. That, however, slows down the communication, and reduces the capac-
ity of the environment as a communication system. An additional drawback of
simply broadcasting communication signals from the base station is the obvious
lack of interception security. A different user of the system who also knows the
propagation behaviour of signals in the environment, can equally well resolve
the series of signals and decode them.
Several approaches have been suggested to circumvent the above mentioned
drawbacks of communication in multiple-scattering environments. Some very
promising techniques are based on the time-reversibility property of propagat-
ing wave-fields [18, 19, 29, 32, 34, 35, 37, 44]. The basic idea is as follows. The
user who wants to communicate with the base station, starts the communica-
tion process by sending a short pilot signal through the environment. The base
station receives this signal as a long and complex signal due to multipathing.
It time-reverses the received signal and sends it back into the environment.
The backpropagating waves will produce a complicated wave-field everywhere
due to the many interfering parts of the emitted signal. However, due to the
time-reversibility of the wave-fields, one expects that the interference will be
constructive at the position of the user who sent the original pilot signal, and
mainly destructive at all other positions. Therefore, the user will receive a short
signal very similar to (ideally, a time-reversed replica of) the originally sent pilot
signal. All other users who might be in the environment at the same time will
only receive noise speckle due to incoherently interfering contributions of the
backpropagating field. If the base station sends the individual elements (‘ones’
and ‘zeros’) of the intended message in a phase-encoded form as a long overlap-
ping string of signals, the superposition principle will ensure that, at the user
position, this string of signals will appear as a series of short well-separated sig-
nals, each resembling some phase-shifted (and time-reversed) form of the pilot
signal.
In order to find out whether this theoretically predicted scenario actually
takes place in a real multiple-scattering environment like the ocean, Kuperman
et al. have performed a series of four experiments in a shallow ocean environment
between 1996 and 2000, essentially following the above described scenario. The
experiments have been performed at a Mediterranean location close to the Italian
coast. (A similar setup was also used in an experiment performed in 2001 off the
coast of New England which has been reported in Yang [49].) A schematic view
of these experiments is shown in Fig. 1. The single ‘user’ is replaced here by a
‘probe source’, and the ‘source-receive array’ (SRA) plays the role of the ‘base
station’. An additional vertical receive array (VRA) was deployed at the position
of the probe source in order to measure the temporal and spatial spread of the
backpropagating fields in the neighbourhood of the probe source location. In this
shallow environment (depths of about 100–200 m, and distances between 10 and
30 km) the multipathing of the waves is mostly caused by multiple reflections
at the surface and the bottom of the ocean. The results of the experiments
have been reported in [18, 26, 31]. They show that in fact a strong spatial and
temporal focusing of the backpropagating waves occurs at the source position.
Time-Reversal and the Adjoint Imaging Method in Communication 139
~ 130 m
~ 10 km
U3 A4
A7
U5
scattering
U7
environment
(especially bandwidth) which are available for this technology [19, 29, 32, 37].
One big advantage of time-reversal techniques is that they are automatically
adapted to the complex environment and that they can be very fast since they
do not require heavy signal processing at the receiver or user side. In [34, 35],
an iterated time-reversal scheme for the optimal refocusing of signals in such
a MIMO communication system was proposed, which we will describe in more
details in Sect. 13.
We will establish a direct link between the time-reversal technique and so-
lution strategies for inverse problems. As an application of this relationship, we
will derive iterative time-reversal schemes for the optimization of wireless or un-
derwater acoustic MIMO communication systems. The derivation is performed
completely in time-domain, for very general first order symmetric hyperbolic sys-
tems describing wave propagation phenomena in a complex environment. One
of the schemes which we derive, in a certain sense the ‘basic one’, will turn out
to be practically equivalent to the scheme introduced in [34, 35], although the
derivation uses different tools. Therefore, we provide a new interpretation of that
scheme. The other schemes which we introduce are new in this application, and
can be considered as either generalizations of the basic scheme, or as independent
alternatives to that scheme. Each of them addresses slightly different objectives
and has its own very specific characteristics.
u(x, 0) = 0. (2)
Time-Reversal and the Adjoint Imaging Method in Communication 141
We mention that the dissipative case Φ(x) = 0 can be treated as well in our
framework, and yields analogous results to those presented here.
with ⎛ ⎞
0 −∂3 ∂2
⎜ ⎟
⎜ ⎟
Ξ = ⎜ ∂3 0 −∂1 ⎟ .
⎝ ⎠
−∂2 ∂1 0
As already mentioned, also elastic waves can be treated in the general framework
of symmetric hyperbolic systems. For more details we refer to [33, 41].
144 O. Dorn
s = (s1 , . . . , sJ ) ∈ Ẑ
r = (r1 , . . . , rK ) ∈ Z
with J K
Ẑ = L2 ([0, T ])N , Z = L2 ([0, T ])N .
The two signal spaces Z and Ẑ introduced above are equipped with the inner
products
J
(1) (2) (1) (2)
s ,s = sj (t), sj (t) dt
Ẑ [0,T ] N
j=1
K
(1) (2) (1) (2)
r ,r = rk (t), rk (t) dt.
Z [0,T ] N
k=1
Each user Uj and each antenna Ak at the base station can send a given
signal sj (t) or rk (t), respectively. This gives rise to a source distribution q̂j (x, t),
j = 1, . . . , J, or qk (x, t), k = 1, . . . , K, respectively. Here and in the following we
will use in our notation the following convention. If one symbol appears in both
forms, with and without a ‘hat’ (ˆ) on top of this symbol, then all quantities with
the ‘hat’ symbol are related to the users, and those without the ‘hat’ symbol to
the antennas at the base station.
Each of the sources created by a user or by a base antenna will appear on
the right hand side of (1) as a mathematical source function and gives rise to a
corresponding wave field which satisfies (1), (2). When solving the system (1),
(2), typically certain Sobolev spaces need to be employed for the appropriate
description of the underlying function spaces (see for example [11]). For our
Time-Reversal and the Adjoint Imaging Method in Communication 145
and which we have equipped with the usual energy inner product
u(t), v(t)U = Γ (x)u(x, t), v(x, t)N dxdt,
[0,T ] Ω
u2U = u , uU .
Also here, in order to simplify the notation, we will use the same space when
considering functions in the shifted time interval [T, 2T ] instead of [0, T ].
Typically, when a user or an antenna at the base station transforms a signal
into a source distribution, it is done according to a very specific antenna charac-
teristic which takes into account the spatial extension of the user or the antenna.
We will model this characteristic at the user by the functions γ̂j (x), j = 1, . . . , J,
and for base antennas by the functions γk (x), k = 1, . . . , K. With these func-
tions, we can introduce the linear ‘source operators’ Q̂ and Q mapping signals
s at the set of users and r at the set of base antennas into the corresponding
source distributions q̂(x, t) and q(x, t), respectively. They are given as
J
Q̂ : Ẑ → U, q̂(x, t) = Q̂s = γ̂j (x)sj (t),
j=1
K
Q : Z → U, q(x, t) = Qr = γk (x)rk (t).
k=1
We will assume that the functions γ̂j (x) are supported on a small neighbour-
hood V̂j of the user location d̂j , and that the functions γk (x) are supported on
a small neighbourhood Vk of the antenna location dk . Moreover, all these neigh-
bourhoods are strictly disjoint to each other. For example, the functions γ̂j (x)
could be assumed to be L2 -approximations of the Dirac delta measure δ(x − d̂j )
concentrated at the user locations d̂j , and the functions γk (x) could be assumed
to be L2 -approximations of the Dirac delta measure δ(x − dk ) concentrated at
the antenna locations dk .
Both, users and base antennas can also record incoming fields u ∈ U and
transform the recorded information into signals. Also here, this is usually done
according to very specific antenna characteristics of each user and each base
antenna. For simplicity (and without loss of generality), we will assume that
the antenna characteristic of a user or base antenna for receiving signals is the
same as for transmitting signals, namely γ̂j (x) for the user and γk (x) for a base
146 O. Dorn
antenna. (The case of more general source and measurement operators is dis-
cussed in Sect. 16.) With this, we can define the linear ‘measurement operators’
M̂ : U → Ẑ and M : U → Z, respectively, which transform incoming fields
into measured signals, by
sj (t) = (M̂ u)j = γ̂j (x)u(x, t)dx, (j = 1, . . . , J)
Ω
rk (t) = (M u)k = γk (x)u(x, t)dx, (k = 1, . . . , K)
Ω
F : U → U, F q = u,
where u solves the problem (1), (2). As already mentioned, we assume that the
domain Ω is chosen sufficiently large and that the boundary ∂Ω is sufficiently
far away from the users and base antennas, such that there is no energy reaching
the boundary in the time interval [0, T ] (or [T, 2T ]) due to the finite speed of
signal propagation. Therefore, the operator F is well-defined.
Formally, we can now introduce the two linear communication operators A
and B. They are defined as
A : Z → Ẑ, Ar = M̂ F Qr,
B : Ẑ → Z, Bs = M F Q̂s.
The operator A models the following situation. The base station emits the signal
r(t) which propagates through the complex environment. The users measure the
arriving wave fields and transform them into measurement signals. The measured
signals at the set of all users is s(t) = Ar(t). The operator B describes exactly
the reversed situation. All users emit together the set of signals s(t), which
propagate through the given complex environment and are received by the base
station. The corresponding set of measured signals at all antennas of the base
station is just r(t) = Bs(t). No time-reversal is involved so far.
the base station wants that at the other users Uj , j > 1, as little energy as pos-
sible arrives when communicating with the specified user U1 . This is also in the
interest of the other users, who want to use a different ‘channel’ for communicat-
ing at the same time with the base antenna, and want to minimize interference
with the communication initiated by user U1 . The complex environment itself
in which the communication takes place (i.e. the ‘channel’) is assumed to be
unknown to all users and to the base station.
In order to arrive at a mathematical description of this problem, we define
the ‘ideal signal’ s̃(t) received by all users as
Each user only knows its own component of this signal, and the base antenna
does not need to know any component of this ideal signal at all.
Definition 7.1. The inverse problem of communication: In the terminol-
ogy of inverse problems, the above described scenario defines an inverse source
problem, which we call for the purpose of these lecture notes the ‘inverse problem
of communication’. The goal is to find a ‘source distribution’ r̃(t) at the base
station which satisfies the ‘data’ s̃(t) at the users:
The ‘state equation’ relating sources to data is given by the symmetric hyperbolic
system (1), (2).
Remark 7.1. Notice that the basic operator A in (12) is unknown to the users
and the base station since they typically do not know the complicated medium
in which the waves propagate. If the operator A (together with s̃) would be
known at the base station by some means, the inverse source problem formu-
lated above could be solved using classical inverse problems techniques, which
would be computationally expensive but in principle doable. In the given situ-
ation, the user and the base station are able to do physical experiments, which
amounts to ‘applying’ the communication operator A to a given signal. Deter-
mining the operator A explicitly by applying it to a set of basis functions of
Z would be possible, but again it would be too expensive. We will show in the
following that, nevertheless, many of the classical solution schemes known from
inverse problems theory can be applied in this situation even without knowing
the operator A explicitly. The basic tool which we will use is a series of time-
reversal experiments, applied to carefully designed signals at the users and the
base station.
Remark 7.2. A practical template for an iterative scheme for finding an optimal
signal at the base station can be envisioned as follows. User U1 starts the com-
(0)
munication process by emitting an initial signal s1 (t) into the complex environ-
ment. This signal, after having propagated through the complex environment,
finally arrives at the base station and is received there usually as a relatively long
and complicated signal due to the multiple scattering events it experienced on its
148 O. Dorn
way. When the base station receives such a signal, it processes it and sends a new
signal r(1) (t) back through the environment which is received by all users. After
receiving this signal, all users become active. The user U1 compares the received
signal with the pilot signal. If the match is not good enough, the received signal
is processed by this user in order to optimize the match with the pilot signal
when receiving the next iterate from the base station. All other users identify
the received signal as unwanted noise, and process it with the goal to receive
in the next iterate from the base station a signal with lower amplitude, such
that it does not interfere with their own communications. All users send now
their processed signals, which together define s(1) (t), back to the base station.
The base station receives them all simultaneously, again usually as a long and
complicated signal, processes this signal and sends a new signal r(2) (t) back into
the environment which ideally will match the desired signals at all users better
than the previously emitted signal r(1) (t). This iteration stops when all users
are satisfied, i.e. when user U1 receives a signal which is sufficiently close to the
pilot signal α(t), and the energy or amplitude of the signals arriving at the other
users has decreased enough in order not to disturb their own communications.
After this learning process of the channel has been completed, the user U1 can
now start communicating safely with the base antenna using the chosen pilot
signal α(t) for decoding the received signals.
Similar schemes have been suggested in [18, 19, 29, 32, 37, 44] in a single-step
fashion, and in [34, 35] performing multiple steps of the iteration. The main
questions to be answered are certainly which signals each user and each base
antenna needs to emit in each step, how these signals need to be processed,
at which stage this iteration should be terminated, and which optimal solution
this scheme is expected to converge to. One of the main objectives of these
lecture notes is to provide a theoretical framework for answering these questions
by combining basic concepts of inverse problems theory with experimental time-
reversal techniques.
In the basic approach we propose to use the gradient method for finding the
minimum of (14). In this method, in each iteration a correction δr is sought for
Time-Reversal and the Adjoint Imaging Method in Communication 149
a guess r which points into the negative gradient direction −A∗ (Ar − s̃) of the
cost functional (14). In other words, starting with the initial guess r(0) = 0, the
iteration of the gradient method goes as follows:
r(0) = 0
(15)
r(n+1) = r(n) − β (n) A∗ (Ar(n) − s̃),
where β (n) is the step-size at iteration number n. Notice that the signals r(n) are
measured at the base station, whereas the difference Ar(n) − s̃ is determined at
the users. In particular, s̃ is the pilot signal only known by the user who defined
it, combined with zero signals at the remaining users. Ar(n) is the signal received
by the users at the n-th iteration step.
M ∗ = Γ −1 Q, M̂ ∗ = Γ −1 Q̂,
(16)
Q∗ = M Γ, Q̂∗ = M̂ Γ.
z(x, T ) = 0, (18)
z(x, t) = 0 on ∂Ω × [0, T ]. (19)
Then
F ∗ v = Γ −1 (x)z(x, t). (20)
Proof: The proof is given in Appendix B.
Remark 9.1. This procedural characterization of the adjoint operator is often
used in solution strategies of large scale inverse problems, where it naturally leads
to so-called ‘backpropagation strategies’. See for example [13, 14, 25, 27, 36, 48]
and the references given there.
150 O. Dorn
Remark 9.2. Notice that in the adjoint system (17)–(19) ‘final value conditions’
are given at t = T in contrast to (1)–(4) where ‘initial value conditions’ are
prescribed at t = 0. This corresponds to the fact that time is running backward
in (17)–(19) and forward in (1)–(4).
Theorem 10.1. Let (φ, ψ)T ∈ U and let (va , pa )T be the solution of the adjoint
system
∂va
−ρ − gradpa (x, t) = ρ(x)φ(x, t) (24)
∂t
∂pa
−κ − divva (x, t) = κ(x)ψ(x, t) (25)
∂t
va (x, T ) = 0, pa (x, T ) = 0 in Ω, (26)
with t ∈ [0, T ] and zero boundary conditions at ∂Ω × [0, T ]. Then we have
−1
∗ φ ρ va (x, t)
F = . (27)
ψ κ−1 pa (x, t)
Proof: This theorem is just an application of Theorem 9.2 to the acoustic sym-
metric hyperbolic system. For the convenience of the reader, we will give a direct
proof as well in Appendix C.
M Sa = Ta M, Sa Q = QTa ,
(30)
M̂ Sa = T̂a M̂ , Sa Q̂ = Q̂T̂a .
∂
−ρ v̂tr (x, τ ) − gradp̂tr (x, τ ) = qv (x, τ ) (37)
∂τ
∂
−κ p̂tr (x, τ ) − divv̂tr (x, τ ) = qp (x, τ ) (38)
∂τ
v̂tr (x, T ) = 0, p̂tr (x, T ) = 0 in Ω. (39)
Taking into account the definition of qv and qp , we see that
v̂tr va
=
p̂tr pa
Theorem 11.1. Let (φ, ψ)T ∈ U and let (Ea , Ha )T be the solution of the adjoint
system
∂Ea
− + curlHa (x, t) + σEa = (x)φ(x, t) (43)
∂t
∂Ha
−µ − curlEa (x, t) = µ(x)ψ(x, t) (44)
∂t
Time-Reversal and the Adjoint Imaging Method in Communication 153
M Se = Te M, Se Q = QTe ,
(49)
M̂ Se = T̂e M̂ , Se Q̂ = Q̂T̂e .
M S = T M, SQ = QT ,
(59)
M̂ S = T̂ M̂ , S Q̂ = Q̂T̂ .
With this, we can prove the following theorem which provides the fundamental
link between time-reversal and inverse problems.
Theorem 12.2. We have
A∗ = T B T̂ . (61)
Proof: Recall that the adjoint operator A∗ can be decomposed as A∗ =
Q∗ F ∗ M̂ ∗ . With Theorem 9.1, Theorem 12.1, and Lemma 12.1, it follows there-
fore that
A∗ = M Γ Γ −1 SF SΓ Γ −1 Q̂
= M SF S Q̂
= T M F Q̂T̂
= T B T̂ ,
Remark 12.1. The above theorem provides a direct link between the adjoint op-
erator A∗ , which plays a central role in the theory of inverse problems, and a
physical experiment modelled by B. The expression T B T̂ defines a ‘time-reversal
experiment’. We will demonstrate in the following sections how we can make use
of this relationship in order to solve the inverse problem of communication by a
series of physical time-reversal experiments.
Remark 12.2. We mention that the above results hold as well for elastic waves
with a suitable definition of the elastic time-reversal mirrors. We leave out the
details for brevity.
156 O. Dorn
The results achieved above give rise to the following experimental procedure for
applying the gradient method (15) to the inverse problem of communication as
formulated in Sects. 7 and 8. First, the pilot signal s̃(t) is defined by user U1 as
described in (11). Moreover, we assume that the first guess r(0) (t) at the base
station is chosen to be zero. Then, using Theorem 12.2, we can write the gradient
method (15) in the equivalent form
r(0) = 0
(62)
r(n+1) = r(n) + β (n) T B T̂ (s̃ − Ar(n) ),
Remark 13.2. We mention that several refinements of this scheme are possible
and straightforward. For example, a weighted inner product can be introduced
for the user signal space Ẑ which puts different preferences on the satisfaction of
the user objectives during the iterative optimization process. For example, if the
‘importance’ of suppressing interferences with other users is valued higher than
to get an optimal signal quality at the specified user U1 , a higher weight can be
put into the inner product at those users which did not start the communication
process. A user who does not care about these interferences, simply puts a very
small weight into his component of the inner product of Ẑ.
Remark 13.3. Notice that there is no mechanism directly built into this proce-
dure which prevents the energy emitted by the base antenna to increase more
than the communication system can support. For example, if the subspace of
signals
Z0 := {r(t) : Ar = 0}
is not empty, then it might happen that during the iteration described above (e.g.
due to noise) an increasing amount of energy is put into signals emitted by the
base station which are in this subspace and which all produce zero contributions
to the measurements at all users. More generally, elements of the subspace of
signals
Zε := {r(t) : ArẐ < εrZ },
for a very small threshold 0 < << 1, might cause problems during the iteration
if the pilot signal s̃(t) chosen by the user has contributions in the subspace
AZε (i.e. in the space of all s = Ar with r ∈ Zε ). This is so because in the
effort of decreasing the mismatch between Ar and s̃(t), the base antenna might
need to put signals with high energy into the system in order to get only small
improvements in the signal match at the user side. Since the environment (and
therefore the operator A) is unknown a-priori, it is difficult to avoid the existence
of such contributions in the pilot signal.
One possible way to prevent the energy emitted by the base station to increase
artificially would be to project the signals r(n) (t) onto the orthogonal comple-
ments of the subspaces Z0 or Zε (if they are known or can be constructed by
some means) prior to their emission. Alternatively, the iteration can be stopped
at an early stage before these unwanted contributions start to build up. (This
in fact has been suggested in [34, 35]).
In the following subsection we introduce an alternative way of ensuring that
the energy emitted by the base station stays reasonably bounded in the effort of
fitting the pilot signal at the users.
158 O. Dorn
1 λ
J˜(r) = Ar − s̃ , Ar − s̃ + r, r (65)
2 Ẑ 2 Z
The negative gradient direction is now given by −A∗ (Ar − s̃) − λr, such that the
regularized iteration reads:
r(0) = 0
(66)
r(n+1) = r(n) + β (n) T B T̂ (s̃ − Ar(n) ) − λr(n) ,
r(0) = 0
s(n) = s̃ − Ar(n) (67)
r(n+1) = r(n) + β (n) T B T̂ s(n) − β (n) λr(n) .
Comparing with (63), we see that the adaptations which need to be applied in
the practical implementation for stabilizing the basic algorithm can easily be
done.
In this section we want to propose an alternative scheme for solving the inverse
problem of communication. As mentioned above, a major drawback of the ba-
sic approach (13) is that the energy emitted by the base station is not limited
explicitly when solving the optimization problem. The regularized version pre-
sented above alleviates this problem. However, we want to mention here that,
under certain assumptions, there is an alternative scheme which can be em-
ployed instead and which has an energy constraint directly built in. Under the
formal assumption that there exists at least one (and presumably more than one)
Time-Reversal and the Adjoint Imaging Method in Communication 159
Here, the operator (AA∗ )−1 acts as a filter on the pilot signal s̃. Instead of
sending the pilot signal to the base station, the users send the filtered version
of it. Certainly, a method must be found in order to apply the filter (AA∗ )−1
to the pilot signal. One possibility of doing so would be to try to determine the
operator AA∗ explicitly by a series of time-reversal experiments on some set of
basis functions of Ẑ, and then invert this operator numerically. However, this
might not be practical in many situations. It certainly would be slow and it
would involve a significant amount of signal-processing. Therefore, we propose
an alternative procedure. First, we notice that there is no need to determine the
whole operator (AA∗ )−1 , but that we only have to apply it to one specific signal,
namely s̃. Let us introduce the short notation
C = AA∗ .
In this notation, we are looking for a signal ŝ ∈ Ẑ such that Cŝ = s̃. We propose
to solve this equation in the least squares sense:
where C ∗ is the adjoint operator to C and β (n) is again some step-size. Expanding
this expression, and taking into account C ∗ = C and A∗ = T B T̂ , we arrive at
s(0) = 0
(72)
s(n+1) = s(n) + β (n) AT B T̂ s̃ − AT B T̂ s(n) .
where IẐ denotes the identity operator in Ẑ and λ > 0 is some suitably chosen
regularization parameter. The necessary adjustments in the gradient iteration
for applying (AA∗ + λIẐ )−1 to s̃ are easily done. We only mention here the
resulting procedure for the implementation of this gradient method by a series
of time-reversal experiments:
Again, the last step shown above is a final application of A∗ to the result of
the gradient iteration for calculating ŝ = (AA∗ + λIẐ )−1 s̃, which yields then
rM N r . This will then be the signal to be applied by the base station during the
communication process with the user U1 .
We have introduced above the regularized least squares solution of the inverse
problem of communication, namely
rLSr = Minr Ar − s̃2Ẑ + λr2Z (76)
with λ > 0 being the regularization parameter. In Hilbert spaces, the solution
of (76) has an explicit form. It is
C = A∗ A + λIZ ,
r(0) = 0
(79)
r(n+1) = r(n) + β (n) (T B T̂ A + λIZ ) r̃ − (T B T̂ A + λIZ )r(n) .
where rν (t) appears on the right hand side at the ν-th position. Our new mea-
surement operator Mν , and the new source operator Qν , are then defined by
Time-Reversal and the Adjoint Imaging Method in Communication 163
Mν : U → Y, Mν u = Pν M u
(81)
Qν : Y → U, Qν rν = QPν∗ rν .
Analogous definitions are done for Ŷ , P̂ν , M̂ν and Q̂ν at the users.
Obviously, we will have to replace now in the above derivation of the iter-
ative time-reversal procedure all measurement operators M by Mν (and M̂ by
M̂ν ) and all source operators Q by Qν (and Q̂ by Q̂ν ). In particular, the new
‘communication operators’ are now given by
Aν : Y → Ŷ , Aν rν = M̂ν F Qν rν ,
(82)
Bν : Ŷ → Y, Bν sν = Mν F Q̂ν sν .
In the following two theorems we show that the main results presented so far
carry over to these newly defined operators.
Proof: The proof is an easy exercise using (81) and Theorem 9.1.
Theorem 16.2. It is
A∗ν = T Bν T̂ . (84)
Proof: The proof is now identical to the proof of Theorem 12.2, using
Theorem 16.1 instead of Theorem 9.1.
Remark 16.1. In fact, it is easy to verify that all results of these notes remain
valid for arbitrarily defined linear measurement operators
MU : U → ZU , MA : U → ZA ,
where ZU and ZA are any meaningful signal spaces at the users and the base
antennas, respectively. The only requirement is that it is experimentally possible
to apply signals according to the source operators defined by
QU : ZU → U, QU = Γ MU∗
∗
QA : ZA → U, QA = Γ MA
à : ZA → ZU , Ãr = MU F QA r,
B̃ : ZU → ZA , B̃s = MA F QU s.
Now the proof to Theorem 12.2 directly carries over to this generalized situation,
such that we have also here
Ã∗ = T B̃ T̂ .
This yields iterative time-reversal schemes completely analogous to those pre-
sented above.
Acknowledgments
These lecture notes have been written during the stay of the author at L2S-
Supélec, France, in the spring of 2004, and are inspired by many discussions
Time-Reversal and the Adjoint Imaging Method in Communication 165
Therefore,
M u, rZ = u, M ∗ rU
with
K
M ∗ r = Γ −1 (x) γk (x)r(t) = Γ −1 (x)Qr.
k=1
M ∗ = Γ −1 Q, M̂ ∗ = Γ −1 Q̂. (85)
Q∗ = M Γ, Q̂∗ = M̂ Γ
where n(x) = (ν1 (x), ν2 (x), ν3 (x)) is the outward normal at ∂Ω in the point x.
Notice that we have augmented Green’s formula in (86) by some terms which
appear in identical form on the left hand side and on the right hand side.
We will assume here that the boundary is far away from the sources and
receivers and that no energy enters Ω from the outside, such that during the
time interval of interest [0, T ] all fields along this boundary are identically zero.
This is expressed by the boundary conditions given in (4) and (19). Let u(x, t) be
a solution of (1), (2), (4), and z(x, t) a solution of (17)–(19). Then the first term
on the left hand side of (86) and the third term on the right hand side cancel
each other because of (1). The second term on the left hand side and the first
term on the right hand side cancel each other because of (17). The (t = T )-term
and the (t = 0) term vanish due to (18) and (2), respectively, and the boundary
integral vanishes because of (4) and (19). The remaining terms (i.e. the third
term on the left hand side and the second term on the right hand side) can be
written as
F q, vU = q, F ∗ vU ,
with F ∗ v = Γ −1 (x)z(x, t) as defined in (20).
This equation has the form (86). Notice that we have augmented Green’s formula
in (87), as already shown in (86), by some terms which appear in identical form
on the left hand side and on the right hand side.
The first term on the left hand side of (87) and the third term on the right
hand side cancel each other due to (21), (22). The second term on the left
hand side and the first term on the right hand side cancel each other because
of (24), (25). The (t = T )-terms and the (t = 0)-terms vanish due to (26),
(23), respectively, and the boundary terms vanish because of zero boundary
conditions. We are left over with the equation
T T
[ρvf φ + κpf ψ] dxdt = [qv va + qp pa ] dxdt
0 Ω 0 Ω
References
1. Bal G and Ryzhik L 2003 Time reversal and refocusing in random media SIAM
J. Appl. Math. 63 1475–98
2. Bardos C and Fink M 2002 Mathematical foundations of the time reversal mirror
Asymptotic Analysis 29 157–182
3. Berryman J G, Borcea L, Papanicolaou G C and Tsogka C 2002 Statistically stable
ultrasonic imaging in random media Journal of the Acoustic Society of America
112 1509–1522
4. Bleistein N, Cohen J K, and Cohen J W 2001 Mathematics of multidimensional
seismic imaging, migration, and inversion (Springer: New York)
5. Blomgren P, Papanicolaou G, and Zhao H 2002 Super-Resolution in Time-Reversal
Acoustics, Journal of the Acoustical Society of America 111 238–248
6. Borcea L, Papanicolaou G and Tsogka C 2003 Theory and applications of time
reversal and interferometric imaging Inverse Problems 19 5139–5164
7. Cheney M, Isaacson D and Lassas M 2001 Optimal Acoustic Measurements SIAM
J Appl Math, 61 1628–1647
8. Claerbout J 1985 Fundamentals of geophysical data processing: with applications
to petroleum prospecting, (CA: Blackwell Scientific Publications, Palo Alto)
9. Collins M D and Kuperman W A 1991 Focalization: Environmental focusing and
source localization J. Acoust. Soc. Am. 90 (3) 1410–1422
10. Collins M D and Kuperman W A 1994 Inverse problems in ocean acoustics Inverse
Problems 10 1023–1040
Time-Reversal and the Adjoint Imaging Method in Communication 169
11. Dautray R and Lions J L Mathematical Analysis and Numerical Methods for Sci-
ence and Technology Vol. 1–6, Springer.
12. Devaney A 2004 Superresolution processing of multi-static data using time reversal
and MUSIC to appear in J. Acoust. Soc. Am.
13. Dorn O 1998 A transport-backtransport method for optical tomography Inverse
Problems 14 1107–1130
14. Dorn O, Bertete-Aguirre H, Berryman J G and Papanicolaou G C 1999 A nonlin-
ear inversion method for 3D electromagnetic imaging using adjoint fields Inverse
Problems 15 1523-1558
15. Dowling D R and Jackson D R 1992 Narrow band performance of phase conjugate
arrays in dynamic random media J. Acoust. Soc. Am. 91 3257-3277
16. Dowling D R 1993 Acoustic pulse compression using passive phase-conjugate
processing J. Acoust. Soc. Am. 95 (3) 1450–1458
17. Draeger C and Fink M 1997 One-Channel Time Reversal of Elastic Waves in a
Chaotic 2D-Silicon Cavity Physical Review Letters 79 (3) 407–410
18. Edelmann G F, Akal T, Hodgkiss W S, Kim S, Kuperman W A and Chung H C
2002 An initial demonstration of underwater acoustic communication using time
reversal IEEE J. Ocean. Eng. 27 (3) 602–609
19. Emami S M, Hansen J, Kim A D, Papanicolaou G, Paulraj A J, Cheung D and
Prettie C 2004 Predicted time reversal performance in wireless communications
using channel measurements submitted
20. Fink M 1997 Time reversed acoustics Phys. Today 50 34-40
21. Fink M and Prada C 2001 Acoustic time-reversal mirrors Inverse Problems 17
R1–R38
22. Foschini G J 1996 Layered space-time architecture for wireless communications in
a fading environment Bell Labs Technical Journal 1 (2) 41–59
23. Fouque J P, Garnier J and Nachbin A 2004 Time reversal for dispersive waves in
random media to appear in SIAM J. Appl. Math.
24. Gesbert D, Shafi M, Shiu D, Smith P J and Naguib A 2003 From Theory to Prac-
tice: An Overview of MIMO Space-Time Coded Wireless Systems IEEE J. Select.
Areas Commun. 21 (3) (tutorial paper) 281–301
25. Haber E, Ascher U and Oldenburg D 2000 On optimization techniques for solving
nonlinear inverse problems Inverse Problems 16 1263–1280
26. Hodgkiss W S, Song H C, Kuperman W A, Akal T, Ferla C and Jackson D R 1999
A long range and variable focus phase conjugation experiment in shallow water
J. Acoust. Soc. Am. 105 1597–1604
27. Hursky P, Porter M B, Cornuelle B D, Hodgkiss W S, and Kuperman W A 2004
Adjoint modeling for acoustic inversion J. Acoust. Soc. Am. 115 (2) 607–619
28. Jackson D R and Dowling D R 1991 Phase conjugation in underwater acoustics J.
Acoust. Soc. Am. 89 (1) 171–181
29. Kim A D, Blomgren P and Papanicolaou G 2004 Spatial focusing and intersymbol
interference in time reversal submitted
30. Klibanov M V and Timonov A 2003 On the mathematical treatment of time re-
versal Inverse Problems 19 1299–1318
31. Kuperman W A, Hodgkiss W S, Song H C, Akal T, Ferla C, and Jackson D R
1998 Phase conjugation in the ocean: Experimental demonstration of an acoustic
time reversal mirror J. Acoust. Soc. Am. 103 25-40
32. Kyritsi P, Papanicolaou G, Eggers P and Oprea A 2004 MISO time reversal and
delay spread compression for FWA channels at 5GHz preprint
33. Marsden J E and Hughes T J R Mathematical Foundations of Elasticity (Prentice
Hall, 1968)
170 O. Dorn
1 Introduction
In this lecture we shall review some basic facts on the so-called point interactions,
i.e. perturbations of the free Laplacian in Rd , d = 1, 2, 3, supported by a finite set
of points (for a comprehensive treatment we refer to the monograph [AGH-KH]).
At a formal level the operator can be written as
n
“ H = −∆ + αj δyj ” (1)
j=1
In particular one can also consider applications to the heat and the wave
equations.
Physical motivations for the introduction of point interactions were given
by Fermi ([F]) in the analysis of scattering of slow neutrons from a target of
condensed matter.
In fact, for a sufficiently slow neutron, the wavelength is much larger than
the effective range of the nuclear force acting between the neutron and each
nucleus of the target and, on the other hand, such force is extremely intense.
As a consequence it appears reasonable to modelize the interaction between the
neutron and the nucleus by a zero-range potential placed at the position of the
nucleus.
As a first approximation one considers each nucleus in a fixed position yj and
this leads to the basic model Hamiltonian (1) and to the corresponding linear
evolution for the wave function of the neutrons solving the Schrödinger equation.
The main interest of point interactions consists in the fact that they define
simple but non trivial models of short range interactions which are, in fact,
explicitly solvable.
More precisely this means that detailed information about the spectrum and
the eigenfunctions of (1) are available and then all the physical relevant quantities
related to the specific problem analyzed can be explicitly computed.
This fact makes point interactions especially useful in applications where
they can be considered as a first step in a more detailed analysis.
More refined evolution models can be obtained considering the strength
and/or the position of the point interactions as given functions of time (lin-
ear non-autonomous evolution problems) or the strength as a given function of
the solution itself (non linear evolution problems).
From a mathematical point of view, the first step is to give a satisfactory
definition of the formal operator (1) as a selfadjoint operator in L2 (Rd ) corre-
sponding to the intuitive idea of point interactions.
One can start from the reasonable consideration that any rigorous counter-
part of (1) must satisfy Hu = −∆u for any u ∈ C0∞ (Rd \ {y1 , . . . , yn }).
This suggests to define the following restriction of the free Laplacian
It is not hard to see that the operator (2) is symmetric but not selfadjoint
in L2 (Rd ); moreover one selfadjoint extension of (2) is trivial, i.e. it corresponds
to the free Laplacian H0 = −∆, D(H0 ) = H 2 (Rd ).
This naturally leads to the following definition.
and
|(−∆ψ, φ)| ≤ ψH 2 φL2
implying that all functions in H 2 (R3 ) belong to the domain of the operator Ĥ ∗
adjoint to Ĥ.
On the other hand for any z ∈ C \ R+ we have
eik·yi
G̃zi (k) =
k2 − z
In fact the Gzi satisfy, in the sense of distributions, (−∆ − z)Gzi = δyi . Being
the δyi together with their derivatives the only distributions supported by the
discrete set {y1 , . . . , yn } the Gzi and their derivatives are the only distributions
satisfying (3).
We conclude that the eigenspace Nz ⊂ L2 (Rd ) relative to the eigenvalue z
of the adjoint Ĥ ∗ of Ĥ contains all the Gzi and their derivatives as long as they
belong to L2 (Rd ).
174 G. Dell’Antonio et al.
Let us follow the details of this construction in the case of one point interac-
tion in d = 3 placed at the point y ∈ R3 . We recall that for d = 3
√
exp i z|x|
Gz (x) =
4π|x|
D(Hα,y ) =
Φλ (0)
= f ∈ L2 (R3 )|f = Φλ + qGλ (· − y), Φλ ∈ H 2 (R3 ), q= √
α+ λ/4π
(9)
Characterization (9) of the domain implies that the action of (Hα,y + λ)−1 does
not differ from the action of (−∆ + λ)−1 on functions of H 2 (R3 \ {y}) (being
q = 0 in this case).
On the other hand
for any g in the set C0∞ (R3 \ {y}), which is dense in L2 (R3 ). As a consequence,
the function ((Hα,y + λ)−1 − (−∆ + λ)−1 )Gλ belongs to N−λ , which in turn
means that it is proportional to Gλ itself.
The action of the two resolvents then differ only on the one dimensional
subspace generated by Gλ . Being D(Hα,y ) in (9) the range of (Hα,y + λ)−1 one
finally obtains
1
[(Hα,y + λ)−1 g](x) = (Gλ g)(x) + √ Gλ (x − y)(Gλ g)(y) (11)
α+ λ/4π
Notice that the free Laplacian is obtained formally in the limit α → ∞ showing
that one cannot consider α as a coupling constant. Its physical meaning becomes
clear studying scattering theory for point interaction Hamiltonians. In this set-
ting one proves that α is the inverse of the scattering length associated to the
operator Hα,y .
From the explicit expression of the resolvent (11) one can easily deduce the
spectral properties of Hα,y :
– The spectrum of Hα,y is
σ(Hα,y ) = [0, ∞) α ≥ 0
0 1
σ(Hα,y ) = −16π 2 α2 ∪ [0, ∞) α < 0
where the continuous part [0, ∞) is purely absolutely continuous.
– For α < 0 the only eigenvalue is simple and the corresponding normalized
eigenfunction is
√ exp(4πα|x − y|)
ψα (x) = −2α .
|x − y|
A Brief Review on Point Interactions 177
with |k|2 = E.
The explicit form of the spectral decomposition of Hα,y , in terms of the
eigenfunction (in the case α < 0) and of the generalized eigenfunctions, allows
to write the solution of the Schrödinger equation
∂ψt
i = Hα,y ψt , (13)
∂t
corresponding to any initial state in L2 (R3 ), as an integral over the spectral
measure. In fact the integral kernel, in configuration space, for the propagator
of Hα,y can be explicitly computed ([ST]). The same procedure allows to write
the kernel of the semigroup generated by Hα,y ([ABD]).
An implicit characterization of the propagator will be useful in the next
section in order to define time dependent point interactions. A formal inverse
Laplace transform of (11) suggests for the solution of the Schrödinger equation
(13) the following formula showing a free-propagation contribution and a term
representing spherical waves generated at the interaction center
t
ψt (x) = (U (t)ψ0 ) (x) + i ds U (t − s, |x − y|) q(s) (14)
0
where U (t) is the propagator of the free unitary group defined by the kernel
|x−x |2
ei 4t
U (t; x − x ) = e i∆t
(x − x ) =
(4πit)3/2
It is easily checked that (14) is the solution of the Schrödinger equation (13)
corresponding to an initial condition ψ0 ∈ D(Hα,y ) if the function q(t) satisfies
the Volterra integral equation
√ t √ t (U (s)ψ0 ) (y)
q(s)
q(t) + 4 iπα ds √ = 4 iπ ds √ (15)
0 t−s 0 t−s
More precisely one can prove that if q(t) is the unique solution of (15) then (14)
defines a function ψt which for any t ≥ 0 belongs to D(Hα,y ) (q(t) being the
coefficient of the singular part of ψt ) and satisfies (13).
We want to conclude the list of properties of the family of Hamiltonians Hα,y
recalling the expression of the associated quadratic forms.
Quadratic forms are often taken as an alternative way to define point inter-
actions, a way often preferred to the one used above because of their immediate
connection with quantities of physical interest. On D(Hα,y ) the quadratic form
uniquely associated to the operator Hα,y is
178 G. Dell’Antonio et al.
Fα,y (u) =
√
" # λ
= dx |∇(u − qGλ (· − y))| + λ|(u − qGλ (· − y))| − λ|u| + α +
2 2 2
|q|2
R 3 4π
(16)
It is bounded below and closable and defines therefore a Dirichlet form. The
domain of its closure is
Notice that the form domain is significantly larger than H 1 (R3 ). In particular
the form Fα,y is not a small perturbation of the one defined by the Laplacian.
We shall now go back to the n centers case in R3 and summarize the main
results about the selfadjoint extensions of the operator defined in (2).
As it was mentioned at the end of Sect. 1 the dimension of the eigenspaces Nz
relative to the eigenvalue z of the operator adjoint to Ĥ is n for any z ∈ C \ R+
in the case of n centers. Consider any unitary operator V from Nz to Nz̄ and
denote with NzV the subspace of the linear span of functions in Nz and in Nz̄
generated by the linear combinations of the type fz + V fz with fz ∈ Nz . A
computation identical to (7) shows that Ĥ ∗ acts as a symmetric operator only
if its action is restricted to any NzV .
As a consequence in this case selfadjoint extensions of Ĥ are uniquely associ-
ated with the n2 dimensional family of (complex) matrices unitarily connecting
Nz and Nz̄ .
In the literature much attention was given to a particular n dimensional
subfamily of selfadjoint extensions called local. In fact, in analogy with the one
center case, functions in the domain of the n dimensional subfamily of selfadjoint
extensions, corresponding to diagonal unitary matrices V , can be alternatively
characterized by a specific behavior around each interaction center. We briefly
recall the properties of the operators in this subfamily (for an introduction to
non local point interactions see [DG]).
For any α = {α1 , . . . , αn } with αi ∈ R, i = 1, . . . , n and y = {y1 , . . . ,
yn }, yi ∈ R3 , i = 1, . . . , n the operator Hα,y defined by
2 n
D(Hα,y ) = u ∈ L2 (R3 ) | u = φλ + qk Gλ (· − yk )
k=1
n 3
φλ ∈ H (R ),
2 3
φλ (yj ) = [Γα,y (λ)]jk qk , j = 1, ..., n (18)
k=1
∂(rj u)
lim − 4παj (rj u) = 0, j = 1, ..., n (21)
rj →0 ∂rj
This explains the term “local” by which this class of extensions is known; indeed
due to the special form of the matrix (20) the generalized boundary conditions
(21) is placed at each point separately. Any other choice for the symmetric
invertible matrix Γα,y (λ) leads to an extension which is still local in the ter-
minology commonly employed for differential operators (i.e. operators which do
not change the support). The extensions obtained by (20) should be perhaps
termed “strongly local” or “local” with respect to the boundary conditions.
In order to find the explicit expression for the resolvent it is sufficient
to observe that, for any given f ∈ L2 (R3 ), the solution u of the equation
(Hα,y + λ)u = f must be written in the form u = Gλ f + qGλ (· − y) (see
(19)), where the charges q are determined imposing the boundary conditions in
(18) (or equivalently (21)). The result of the easy computation is
n
(Hα,y + λ)−1 = Gλ + [Γα,y (λ)]−1
jk Gλ (· − yj )Gλ (· − yk ) (22)
j,k=1
where λ > 0 is chosen sufficiently large so that the matrix (20) is invertible.
From the analysis of (22) one can derive that the continuous spectrum of
Hα,y is purely absolutely continuous and coincides with the positive real axis.
The point spectrum consists of (at most) n non positive eigenvalues given by
the possible solutions E ≥ 0 of the equation det [Γα,y (−E)] = 0.
Proper and generalized eigenfunctions can be explicitly computed.
Notice that the main tool used in the construction of the selfadjoint exten-
sions Hα,y was the specific behavior of the Green’s function of the Laplacian
at the singularity. Therefore the approach we described above can be adapted
to treat stationary or evolution problems related to perturbations supported by
points of any elliptic operator of the general form
d
∂ ∂
LV = (i + Ak )akj (x)(i + Aj ) + V (x)
∂xk ∂xj
k,j=1
definition in the general case is linked with the possibility of finding an explicit
expression for the Green’s function of the operator LV (as in the case of the
harmonic oscillator).
It is easy to check an additivity property with respect to the potential while
on the other hand, in dimension 2 and 3, point interactions are not additive in
the sense that
LVα,y = LVα∪β,y∪z
β,z
The form domain is then the same of the form domain of the Laplacian and in
fact standard Sobolev inequalities show that Fα,y is a small perturbation of the
form associated to the Laplacian (see ([T]) for details). In terms of quadratic
forms the generalization to n centers placed in {y1 , . . . , yn } ≡ y of strength
{α1 , . . . , αn } ≡ α is immediate
Formula (27) shows that, differently from the higher dimensional cases, in
d = 1 point interactions are genuinely additive and that the dynamical parameters
αi play the role of coupling constants.
A Brief Review on Point Interactions 181
The explicit form of the integral kernel of the resolvent can be computed; in
the single center case one has
√ √ √ √
−1 e− λ|x−x | 2α λ e− λ|x−y| e− λ|x −y|
(Hα,y + λ) (x, x ) = √ − √ √ √
2 λ α+2 λ 2 λ 2 λ
Analyzing the singularities of the resolvent one can easily see that the spec-
trum σ(Hα,y ) of Hα,y has the properties
It is easy to check that in fact (28) and (29) define a unitary flow in L2 (R) whose
generator is Hα,y and therefore they may be used as a definition of Hα,y .
We want briefly to mention the other subfamily of selfadjoint extensions of
Ĥ in d = 1 known as δ interactions.
For each β ∈ R and y ∈ R the operator Hβ,y can be defined via the action
on a function f ∈ L (R) of its resolvent in the following way
2
182 G. Dell’Antonio et al.
2βλ
(Hβ,y + λ)−1 )f = (−∆ + λ)−1 f + √ Gλ (· − y)(Gλ f )(y) (30)
2+β λ
where Gλ is the derivative of the Green’s function for z = −λ, λ > 0
1 √
Gλ = − e− λ|x−y| x > y
2
1 −√λ|x−y|
= e x<y
2
and λ = (2/β)2 if β < 0.
The spectrum of Hβ,y is easily found to be
σ(Hβ,y ) = [0, +∞), β≥0
2 3
4
σ(Hβ,y ) = − 2 ∪ [0, +∞), β<0
β
For further details and properties on this kind of operators see [AGH-KH].
where the potentials Vj can be chosen in L1 (R). Then for any λ > 0 sufficiently
large one has
lim (H + λ)−1 − (Hα,y + λ)−1 = 0 (32)
→0
(
with αj = R dxVj (x) (see [AGH-KH]).
In dimension three the situation is more delicate and the class of approxi-
mating potentials must be properly chosen.
In particular one has to restrict to potentials V such that −∆ + V has a
so-called zero-energy resonance, i.e. there exists a solution ψ ∈ L2loc (R3 ), with
∇ψ ∈ L2 (R3 ), of the equation (−∆ + V )ψ = 0. We call such a solution a zero-
energy resonance function.
A Brief Review on Point Interactions 183
Let us now go back to the second problem mentioned at the beginning of this
section concerning a procedure allowing to approximate (in a way which should
be suitable for applications) a Schrödinger operator with smooth potential via
point interaction Hamiltonians.
The answer is obviously positive in dimension d = 1 for the particular self-
adjoint Hamiltonians we referred to as delta interaction Hamiltonians. In that
184 G. Dell’Antonio et al.
case point interactions are additive and, at least in the sense of quadratic forms,
they correspond to a potential which is a sum of delta functions. It is easy to
prove that any approximation of a potential function through linear combina-
tions of delta measures turns out to be an efficient approximation scheme for
the corresponding operators ([BFT]).
As we mentioned the situation is fairly different in dimension d = 2 and
d = 3. In these cases point interaction Hamiltonians for n centers are in no sense
sum of an unperturbed operator and a potential, as we remarked at the end of
the last section. Moreover point interactions in d = 2 and d = 3 are in no sense
connected to delta measures as we discussed in the first part of this section.
Nevertheless, with some peculiar differences with respect to the one dimen-
sional case, it is possible to work out an approximation procedure that we out-
line here restricting again our attention to the case d = 3. Let us denote with
(N ) (N )
Y N = {y1 , . . . , yN } a configuration of N points in R3 and consider a (prod-
uct) probability measure [ρ(y)dy]⊗N on such configuration space. For simplicity
let us assume that the density is continuous. Let α(y) be a real function on
R3 continuous and bounded away from zero outside a set of Lebesgue measure
(N ) (N )
zero and define αN ≡ {α(y1 ) . . . , α(yN )}. The following theorem was proved
in [FHT]
On a set of configurations of measure increasing to 1 as N tends to infinity,
for any λ positive, large enough
−1 ρ −1
s − lim HN αN ,Y N + λ) = −∆ − + λ (40)
N →∞ α
We want to mention an immediate consequence of the result stated above. Notice
first that any smooth integrable potential V can be written in the form ρ/α
taking
|V (y)| (sgnV )(y)
ρ(y) = ( α= (
R3
|V (y)|dy R3
|V (y)|dy
We conclude that any Schrödinger operator with a smooth integrable potential
can be approximated by point interaction Hamiltonians.
As a final remark we want to stress that (40) shows that positive potentials
are obtained as limits of negative alphas and vice versa. This is a consequence
of the fact that even if all the αi ’s are positive the corresponding n centers point
interaction Hamiltonian can have negative eigenvalues if some, or all, the points
are very close to each other. This depends in turn on the fact that the non
diagonal terms in the matrix (20) can dominate the diagonal terms if there are
couples of very close points.
On the other hand if all the αi ’s are negative the rescaled Hamiltonian shows
N negative eigenvalues all tending to infinity for large N . In this second case
there is no uniform bound from below for the eiganvalues of the operators in the
approximating sequence and in particular no convergence of the corresponding
semigroups can take place.
A Brief Review on Point Interactions 185
∂ψs (t)
i < v(t), > = Bα(t),y(t) (v(t), ψs (t)) ∀v(t) ∈ D(Fα(t),y(t) )
∂t
ψs (s) = f
(44)
186 G. Dell’Antonio et al.
∂η ∂2η
i (t, x) = − 2 (t, x) − |η(t, 0)|2 δ0 η(t, x) (47)
∂t ∂ξ
Since for a class of blow-up solutions ξ(t) of the first equation it has been shown
that |ξ(t)|2 → δ0 , it appears that, for this family of solutions, the second equation
approximates the first one.
Unfortunately, this striking analogy does not hold in higher dimensions.
In the following we are going to consider the three dimensional case. For
details on the one dimensional case see [AT1], [AT2].
Notice that even in the linear three dimensional case the solution of the
evolution problem exhibits a singularity where the interaction is placed. This
means that the nonlinear interaction cannot be introduced as in d = 1. The value
of the function at the origin should be replaced in this case by the coefficient
q(t) of the singular part. This suggests the following formulation of the evolution
problem (see (15))
t
ψt (x) = (U (t)ψ0 )(x) + i dsU (t − s, |x|)q(s)
0
√ t α(s)q(s) √ t (U (s)ψ0 )(0)
q(t) + 4 i π ds √ =4 iπ ds √
0 t−s 0 t−s
α(t) = γ|q(t)|2σ , γ ∈ R, σ > 0
(48)
The natural space where one looks for the solution of a standard nonlinear
Schrödinger equation is H 1 , i.e. the form domain of the generator of the linear
dynamics obtained setting to zero the exponent of the nonlinear term. This
suggests to look for the solution of (48) in the form domain D(Fα,0 ) of Hα,0 . It
is easy to check that the following characterization of D(Fα,0 ) is equivalent to
the one given in (17)
or equivalently
lim |q(t)| = ∞
t→t0
References
[ABD] Albeverio S., Brzeźniak Z., Da̧browski L., Fundamental Solution of the
Heat and Schrödinger Equations with Point Interaction, J. Func. Anal.,
130, 220–254, 1995.
[ADFT1] Adami R., Dell’Antonio G., Figari R., Teta A., The Cauchy Problem for the
Schrödinger Equation in Dimension Three with Concentrated Nonlinearity,
Ann. Inst. H. Poincare’ Non Lineaire, 20, no. 3, 477-500, 2003.
[ADFT2] Adami R., Dell’Antonio G., Figari R., Teta A., Blow Up Solutions for
the Schrödinger Equation with a Concentrated Nonlinearity in Dimension
Three, Ann. Inst. H. Poincare’ Non Lineaire, 21, no. 1, 121-137, 2004.
[AG] Akhiezer N.I., Glazman I.M., Theory of Linear Operators in Hilbert Space,
Vol 2, Pitman, Boston-London-Melbourne, 1981.
[AGH-KH] Albeverio S., Gesztesy F., Högh-Krohn R., Holden H., Solvable Models in
Quantum Mechanics, Springer-Verlag, New York, 1988.
[AT1] Adami R., Teta A., A Simple Model of Concentrated Nonlinearity, in Math-
ematical Results in Quantum Mechanics, Dittrich J., Exner P.,Tater M.
eds., Birkhäuser, 1999.
[AT2] Adami R., Teta A., A Class of Nonlinear Schrödinger Equation with Con-
centrated Nonlinearity, J. Func. An., 180,148-175, 2001.
[BFT] Brasche J.F., Figari R., Teta A., Singular Schrödinger Operator as Limits
of Point Interaction Hamiltonians, Potential Analysis, 8, 163-178, 1998.
[BKB] Bulashenko O.M., Kochelap V.A., Bonilla L.L., Coherent Patterns and
Self-Induced Diffraction of Electrons on a Thin Nonlinear Layer, Phys.
Rev. B, 54, 3, 1996.
[DFT1] Dell’Antonio G.F., Figari R., Teta A., Diffusion of a Particle in Presence
of N Moving Point Sources, Ann. Inst. H. Poincare’ A., 69, 413-424, 1998.
[DFT2] Dell’Antonio G.F., Figari R., Teta A., Schrödinger Equation with Moving
Point Interactions in Three Dimensions, in Stochastic Processes, Physics
and Geometry: New Interplays, vol. I, Gesztesy F., Holden H., Jost J.,
Paycha S., Röckner M., Scarlatti S. eds., A.M.S., Providence, 2000.
[DG] Da̧browski L., Grosse H., On nonlocal point interactions in one, two, and
three dimensions, J. Math. Phys., 26, 2777-2780, 1985.
A Brief Review on Point Interactions 189
[F] Fermi E., Sul moto dei neutroni nelle sostanze idrogenate, Ricerca Sci-
entifica 7, 13-52, 1936 (Engl. Trans. in E. Fermi Collected Papers Vol.1,
University of Chicago, Chicago-London, 1962).
[FHT] Figari R., Holden H., Teta A., A law of large numbers and a central limit
theorem for the Schrödinger operator with zero range potentials, J. Stat.
Phys., 51, 205-214, 1988.
[FT] Figari R., Teta A., A boundary value problem of mixed type on perforated
domains Asymptotic Analysis, 6, 271-284, 1993.
[J-LPS] Jona-Lasinio G., Presilla C., Sjöstrand J., On Schrödinger Equations with
Concentrated Nonlinearities, Ann. Phys., 240, 1-21, 1995.
[KP] Kronig R. de L., Penney W. G., Quantum Mechanics of Electrons in Crystal
Lattices, Proc. Roy. Soc. (London), 130A, 499-513, 1931.
[MA] Malomed B., Azbel M., Modulational Instability of a Wave Scattered by a
Nonlinear Center, Phys. Rev. B, 47, 16, 1993.
[S] Schulman L.S., Application of the propagator for the delta function po-
tential, in Path integrals from mev to Mev, Gutzwiller M.C., Iuomata A.,
Klauder J.K., Streit L. eds., World Scientific, Singapore, 1986.
[ST] Scarlatti S., Teta A., Derivation of the Time-dependent Propagator for the
Three Dimensional Schródinger Equation with One Point Interaction, J.
Phys. A, 23, 1033-1035, 1990.
[Si] Simon B., Quantum Mechanics for Hamiltonians Defined as Quadratic
Forms, Princeton University Press, 1971.
[SY] Sayapova M.R., Yafaev D.R., The evolution operator for time-dependent
potentials of zero radius, Proc. Stek. Inst. Math., 2, 173-180, 1984.
[T] Teta A., Quadratic Forms for Singular Perturbations of the Laplacian, Pub-
lications of the R.I.M.S., Kyoto University, 26, 803-817, 1990.
List of Participants
Published by C.I.M.E
1954 1. Analisi funzionale
2. Quadratura delle superficie e questioni connesse
3. Equazioni differenziali non lineari
1991 111. Topological Methods for Ordinary Differential Equations (LNM 1537)
112. Arithmetic Algebraic Geometry (LNM 1553)
113. Transition to Chaos in Classical and Quantum Mechanics (LNM 1589)
1994 120. Recent Mathematical Methods in Nonlinear Wave Propagation (LNM 1640)
121. Dynamical Systems (LNM 1609)
122. Transcendental Methods in Algebraic Geometry (LNM 1646)
1996 126. Integral Geometry, Radon Transforms and Complex Analysis (LNM 1684)
127. Calculus of Variations and Geometric Evolution Problems (LNM 1713)
128. Financial Mathematics (LNM 1656)
1997 129. Mathematics Inspired by Biology (LNM 1714)
130. Advanced Numerical Approximation of Nonlinear Hyperbolic (LNM 1697)
Equations
131. Arithmetic Theory of Elliptic Curves (LNM 1716)
132. Quantum Cohomology (LNM 1776)
2005 162. Enumerative Invariants in Algebraic Geometry and String Theory to appear
163. Calculus of Variations and Non-linear Partial Differential Equations (LNM 1927)
164. SPDE in Hydrodynamic. Recent Progress and Prospects (LNM 1942)
Vol. 1755: J. Azéma, M. Émery, M. Ledoux, M. Yor Vol. 1780: J. H. Bruinier, Borcherds Products on O(2,1)
(Eds.), Séminaire de Probabilités XXXV (2001) and Chern Classes of Heegner Divisors (2002)
Vol. 1756: P. E. Zhidkov, Korteweg de Vries and Nonlin- Vol. 1781: E. Bolthausen, E. Perkins, A. van der Vaart,
ear Schrödinger Equations: Qualitative Theory (2001) Lectures on Probability Theory and Statistics. Ecole d’
Vol. 1757: R. R. Phelps, Lectures on Choquet’s Theorem Eté de Probabilités de Saint-Flour XXIX-1999. Editor:
(2001) P. Bernard (2002)
Vol. 1758: N. Monod, Continuous Bounded Cohomology Vol. 1782: C.-H. Chu, A. T.-M. Lau, Harmonic Functions
of Locally Compact Groups (2001) on Groups and Fourier Algebras (2002)
Vol. 1759: Y. Abe, K. Kopfermann, Toroidal Groups Vol. 1783: L. Grüne, Asymptotic Behavior of Dynamical
(2001) and Control Systems under Perturbation and Discretiza-
Vol. 1760: D. Filipović, Consistency Problems for Heath- tion (2002)
Jarrow-Morton Interest Rate Models (2001) Vol. 1784: L. H. Eliasson, S. B. Kuksin, S. Marmi, J.-C.
Vol. 1761: C. Adelmann, The Decomposition of Primes in Yoccoz, Dynamical Systems and Small Divisors. Cetraro,
Torsion Point Fields (2001) Italy 1998. Editors: S. Marmi, J.-C. Yoccoz (2002)
Vol. 1762: S. Cerrai, Second Order PDE’s in Finite and Vol. 1785: J. Arias de Reyna, Pointwise Convergence of
Infinite Dimension (2001) Fourier Series (2002)
Vol. 1763: J.-L. Loday, A. Frabetti, F. Chapoton, F. Goi- Vol. 1786: S. D. Cutkosky, Monomialization of Mor-
chot, Dialgebras and Related Operads (2001) phisms from 3-Folds to Surfaces (2002)
Vol. 1764: A. Cannas da Silva, Lectures on Symplectic Vol. 1787: S. Caenepeel, G. Militaru, S. Zhu, Frobenius
Geometry (2001) and Separable Functors for Generalized Module Cate-
Vol. 1765: T. Kerler, V. V. Lyubashenko, Non-Semisimple gories and Nonlinear Equations (2002)
Topological Quantum Field Theories for 3-Manifolds with Vol. 1788: A. Vasil’ev, Moduli of Families of Curves for
Corners (2001) Conformal and Quasiconformal Mappings (2002)
Vol. 1766: H. Hennion, L. Hervé, Limit Theorems for Vol. 1789: Y. Sommerhäuser, Yetter-Drinfel’d Hopf alge-
Markov Chains and Stochastic Properties of Dynamical bras over groups of prime order (2002)
Systems by Quasi-Compactness (2001)
Vol. 1790: X. Zhan, Matrix Inequalities (2002)
Vol. 1767: J. Xiao, Holomorphic Q Classes (2001)
Vol. 1791: M. Knebusch, D. Zhang, Manis Valuations
Vol. 1768: M. J. Pflaum, Analytic and Geometric Study of
and Prüfer Extensions I: A new Chapter in Commutative
Stratified Spaces (2001)
Algebra (2002)
Vol. 1769: M. Alberich-Carramiñana, Geometry of the
Vol. 1792: D. D. Ang, R. Gorenflo, V. K. Le, D. D. Trong,
Plane Cremona Maps (2002)
Moment Theory and Some Inverse Problems in Potential
Vol. 1770: H. Gluesing-Luerssen, Linear Delay-
Theory and Heat Conduction (2002)
Differential Systems with Commensurate Delays: An
Vol. 1793: J. Cortés Monforte, Geometric, Control and
Algebraic Approach (2002)
Numerical Aspects of Nonholonomic Systems (2002)
Vol. 1771: M. Émery, M. Yor (Eds.), Séminaire de Prob-
abilités 1967-1980. A Selection in Martingale Theory Vol. 1794: N. Pytheas Fogg, Substitution in Dynamics,
(2002) Arithmetics and Combinatorics. Editors: V. Berthé, S. Fer-
enczi, C. Mauduit, A. Siegel (2002)
Vol. 1772: F. Burstall, D. Ferus, K. Leschke, F. Pedit,
U. Pinkall, Conformal Geometry of Surfaces in S4 (2002) Vol. 1795: H. Li, Filtered-Graded Transfer in Using Non-
Vol. 1773: Z. Arad, M. Muzychuk, Standard Integral commutative Gröbner Bases (2002)
Table Algebras Generated by a Non-real Element of Small Vol. 1796: J.M. Melenk, hp-Finite Element Methods for
Degree (2002) Singular Perturbations (2002)
Vol. 1774: V. Runde, Lectures on Amenability (2002) Vol. 1797: B. Schmidt, Characters and Cyclotomic Fields
Vol. 1775: W. H. Meeks, A. Ros, H. Rosenberg, The in Finite Geometry (2002)
Global Theory of Minimal Surfaces in Flat Spaces. Vol. 1798: W.M. Oliva, Geometric Mechanics (2002)
Martina Franca 1999. Editor: G. P. Pirola (2002) Vol. 1799: H. Pajot, Analytic Capacity, Rectifiability,
Vol. 1776: K. Behrend, C. Gomez, V. Tarasov, G. Tian, Menger Curvature and the Cauchy Integral (2002)
Quantum Comohology. Cetraro 1997. Editors: P. de Bar- Vol. 1800: O. Gabber, L. Ramero, Almost Ring Theory
tolomeis, B. Dubrovin, C. Reina (2002) (2003)
Vol. 1777: E. García-Río, D. N. Kupeli, R. Vázquez- Vol. 1801: J. Azéma, M. Émery, M. Ledoux, M. Yor
Lorenzo, Osserman Manifolds in Semi-Riemannian (Eds.), Séminaire de Probabilités XXXVI (2003)
Geometry (2002) Vol. 1802: V. Capasso, E. Merzbach, B. G. Ivanoff,
Vol. 1778: H. Kiechle, Theory of K-Loops (2002) M. Dozzi, R. Dalang, T. Mountford, Topics in Spatial
Vol. 1779: I. Chueshov, Monotone Random Systems Stochastic Processes. Martina Franca, Italy 2001. Editor:
(2002) E. Merzbach (2003)
Vol. 1803: G. Dolzmann, Variational Methods for Crys- Vol. 1827: K. M. Pilgrim, Combinations of Complex
talline Microstructure – Analysis and Computation (2003) Dynamical Systems. IX, 118 p, 2003.
Vol. 1804: I. Cherednik, Ya. Markov, R. Howe, G. Lusztig, Vol. 1828: D. J. Green, Gröbner Bases and the Computa-
Iwahori-Hecke Algebras and their Representation Theory. tion of Group Cohomology. XII, 138 p, 2003.
Martina Franca, Italy 1999. Editors: V. Baldoni, D. Bar- Vol. 1829: E. Altman, B. Gaujal, A. Hordijk, Discrete-
basch (2003) Event Control of Stochastic Networks: Multimodularity
Vol. 1805: F. Cao, Geometric Curve Evolution and Image and Regularity. XIV, 313 p, 2003.
Processing (2003) Vol. 1830: M. I. Gil’, Operator Functions and Localization
Vol. 1806: H. Broer, I. Hoveijn. G. Lunther, G. Vegter, of Spectra. XIV, 256 p, 2003.
Bifurcations in Hamiltonian Systems. Computing Singu- Vol. 1831: A. Connes, J. Cuntz, E. Guentner, N. Hig-
larities by Gröbner Bases (2003) son, J. E. Kaminker, Noncommutative Geometry, Martina
Vol. 1807: V. D. Milman, G. Schechtman (Eds.), Geomet- Franca, Italy 2002. Editors: S. Doplicher, L. Longo (2004)
ric Aspects of Functional Analysis. Israel Seminar 2000- Vol. 1832: J. Azéma, M. Émery, M. Ledoux, M. Yor
2002 (2003) (Eds.), Séminaire de Probabilités XXXVII (2003)
Vol. 1808: W. Schindler, Measures with Symmetry Prop- Vol. 1833: D.-Q. Jiang, M. Qian, M.-P. Qian, Mathemati-
erties (2003) cal Theory of Nonequilibrium Steady States. On the Fron-
Vol. 1809: O. Steinbach, Stability Estimates for Hybrid tier of Probability and Dynamical Systems. IX, 280 p,
Coupled Domain Decomposition Methods (2003) 2004.
Vol. 1810: J. Wengenroth, Derived Functors in Functional Vol. 1834: Yo. Yomdin, G. Comte, Tame Geometry with
Analysis (2003) Application in Smooth Analysis. VIII, 186 p, 2004.
Vol. 1811: J. Stevens, Deformations of Singularities Vol. 1835: O.T. Izhboldin, B. Kahn, N.A. Karpenko,
(2003) A. Vishik, Geometric Methods in the Algebraic Theory
Vol. 1812: L. Ambrosio, K. Deckelnick, G. Dziuk, of Quadratic Forms. Summer School, Lens, 2000. Editor:
M. Mimura, V. A. Solonnikov, H. M. Soner, Mathematical J.-P. Tignol (2004)
Aspects of Evolving Interfaces. Madeira, Funchal, Portu- Vol. 1836: C. Nǎstǎsescu, F. Van Oystaeyen, Methods of
gal 2000. Editors: P. Colli, J. F. Rodrigues (2003) Graded Rings. XIII, 304 p, 2004.
Vol. 1813: L. Ambrosio, L. A. Caffarelli, Y. Brenier, Vol. 1837: S. Tavaré, O. Zeitouni, Lectures on Probabil-
G. Buttazzo, C. Villani, Optimal Transportation and its ity Theory and Statistics. Ecole d’Eté de Probabilités de
Applications. Martina Franca, Italy 2001. Editors: L. A. Saint-Flour XXXI-2001. Editor: J. Picard (2004)
Caffarelli, S. Salsa (2003) Vol. 1838: A.J. Ganesh, N.W. O’Connell, D.J. Wischik,
Vol. 1814: P. Bank, F. Baudoin, H. Föllmer, L.C.G. Big Queues. XII, 254 p, 2004.
Rogers, M. Soner, N. Touzi, Paris-Princeton Lectures on Vol. 1839: R. Gohm, Noncommutative Stationary
Mathematical Finance 2002 (2003) Processes. VIII, 170 p, 2004.
Vol. 1815: A. M. Vershik (Ed.), Asymptotic Com- Vol. 1840: B. Tsirelson, W. Werner, Lectures on Probabil-
binatorics with Applications to Mathematical Physics. ity Theory and Statistics. Ecole d’Eté de Probabilités de
St. Petersburg, Russia 2001 (2003) Saint-Flour XXXII-2002. Editor: J. Picard (2004)
Vol. 1816: S. Albeverio, W. Schachermayer, M. Tala- Vol. 1841: W. Reichel, Uniqueness Theorems for Vari-
grand, Lectures on Probability Theory and Statistics. ational Problems by the Method of Transformation
Ecole d’Eté de Probabilités de Saint-Flour XXX-2000. Groups (2004)
Editor: P. Bernard (2003) Vol. 1842: T. Johnsen, A. L. Knutsen, K3 Projective Mod-
Vol. 1817: E. Koelink, W. Van Assche (Eds.), Orthogonal els in Scrolls (2004)
Polynomials and Special Functions. Leuven 2002 (2003) Vol. 1843: B. Jefferies, Spectral Properties of Noncom-
Vol. 1818: M. Bildhauer, Convex Variational Problems muting Operators (2004)
with Linear, nearly Linear and/or Anisotropic Growth Vol. 1844: K.F. Siburg, The Principle of Least Action in
Conditions (2003) Geometry and Dynamics (2004)
Vol. 1819: D. Masser, Yu. V. Nesterenko, H. P. Schlick- Vol. 1845: Min Ho Lee, Mixed Automorphic Forms, Torus
ewei, W. M. Schmidt, M. Waldschmidt, Diophantine Bundles, and Jacobi Forms (2004)
Approximation. Cetraro, Italy 2000. Editors: F. Amoroso, Vol. 1846: H. Ammari, H. Kang, Reconstruction of Small
U. Zannier (2003) Inhomogeneities from Boundary Measurements (2004)
Vol. 1820: F. Hiai, H. Kosaki, Means of Hilbert Space Vol. 1847: T.R. Bielecki, T. Björk, M. Jeanblanc, M.
Operators (2003) Rutkowski, J.A. Scheinkman, W. Xiong, Paris-Princeton
Vol. 1821: S. Teufel, Adiabatic Perturbation Theory in Lectures on Mathematical Finance 2003 (2004)
Quantum Dynamics (2003) Vol. 1848: M. Abate, J. E. Fornaess, X. Huang, J. P. Rosay,
Vol. 1822: S.-N. Chow, R. Conti, R. Johnson, J. Mallet- A. Tumanov, Real Methods in Complex and CR Geom-
Paret, R. Nussbaum, Dynamical Systems. Cetraro, Italy etry, Martina Franca, Italy 2002. Editors: D. Zaitsev, G.
2000. Editors: J. W. Macki, P. Zecca (2003) Zampieri (2004)
Vol. 1823: A. M. Anile, W. Allegretto, C. Ring- Vol. 1849: Martin L. Brown, Heegner Modules and Ellip-
hofer, Mathematical Problems in Semiconductor Physics. tic Curves (2004)
Cetraro, Italy 1998. Editor: A. M. Anile (2003) Vol. 1850: V. D. Milman, G. Schechtman (Eds.), Geomet-
Vol. 1824: J. A. Navarro González, J. B. Sancho de Salas, ric Aspects of Functional Analysis. Israel Seminar 2002-
C ∞ – Differentiable Spaces (2003) 2003 (2004)
Vol. 1825: J. H. Bramble, A. Cohen, W. Dahmen, Mul- Vol. 1851: O. Catoni, Statistical Learning Theory and
tiscale Problems and Methods in Numerical Simulations, Stochastic Optimization (2004)
Martina Franca, Italy 2001. Editor: C. Canuto (2003) Vol. 1852: A.S. Kechris, B.D. Miller, Topics in Orbit
Vol. 1826: K. Dohmen, Improved Bonferroni Inequal- Equivalence (2004)
ities via Abstract Tubes. Inequalities and Identities of Vol. 1853: Ch. Favre, M. Jonsson, The Valuative Tree
Inclusion-Exclusion Type. VIII, 113 p, 2003. (2004)
Vol. 1854: O. Saeki, Topology of Singular Fibers of Dif- Vol. 1878: R. Cerf, The Wulff Crystal in Ising and Percol-
ferential Maps (2004) ation Models, Ecole d’Eté de Probabilités de Saint-Flour
Vol. 1855: G. Da Prato, P.C. Kunstmann, I. Lasiecka, XXXIV-2004. Editor: Jean Picard (2006)
A. Lunardi, R. Schnaubelt, L. Weis, Functional Analytic Vol. 1879: G. Slade, The Lace Expansion and its Applica-
Methods for Evolution Equations. Editors: M. Iannelli, tions, Ecole d’Eté de Probabilités de Saint-Flour XXXIV-
R. Nagel, S. Piazzera (2004) 2004. Editor: Jean Picard (2006)
Vol. 1856: K. Back, T.R. Bielecki, C. Hipp, S. Peng, Vol. 1880: S. Attal, A. Joye, C.-A. Pillet, Open Quantum
W. Schachermayer, Stochastic Methods in Finance, Bres- Systems I, The Hamiltonian Approach (2006)
sanone/Brixen, Italy, 2003. Editors: M. Fritelli, W. Rung- Vol. 1881: S. Attal, A. Joye, C.-A. Pillet, Open Quantum
galdier (2004) Systems II, The Markovian Approach (2006)
Vol. 1857: M. Émery, M. Ledoux, M. Yor (Eds.), Sémi- Vol. 1882: S. Attal, A. Joye, C.-A. Pillet, Open Quantum
naire de Probabilités XXXVIII (2005) Systems III, Recent Developments (2006)
Vol. 1858: A.S. Cherny, H.-J. Engelbert, Singular Stochas- Vol. 1883: W. Van Assche, F. Marcellàn (Eds.), Orthogo-
tic Differential Equations (2005) nal Polynomials and Special Functions, Computation and
Application (2006)
Vol. 1859: E. Letellier, Fourier Transforms of Invariant
Functions on Finite Reductive Lie Algebras (2005) Vol. 1884: N. Hayashi, E.I. Kaikina, P.I. Naumkin,
I.A. Shishmarev, Asymptotics for Dissipative Nonlinear
Vol. 1860: A. Borisyuk, G.B. Ermentrout, A. Friedman,
Equations (2006)
D. Terman, Tutorials in Mathematical Biosciences I.
Vol. 1885: A. Telcs, The Art of Random Walks (2006)
Mathematical Neurosciences (2005)
Vol. 1886: S. Takamura, Splitting Deformations of Dege-
Vol. 1861: G. Benettin, J. Henrard, S. Kuksin, Hamil- nerations of Complex Curves (2006)
tonian Dynamics – Theory and Applications, Cetraro, Vol. 1887: K. Habermann, L. Habermann, Introduction to
Italy, 1999. Editor: A. Giorgilli (2005) Symplectic Dirac Operators (2006)
Vol. 1862: B. Helffer, F. Nier, Hypoelliptic Estimates and Vol. 1888: J. van der Hoeven, Transseries and Real Differ-
Spectral Theory for Fokker-Planck Operators and Witten ential Algebra (2006)
Laplacians (2005) Vol. 1889: G. Osipenko, Dynamical Systems, Graphs, and
Vol. 1863: H. Führ, Abstract Harmonic Analysis of Con- Algorithms (2006)
tinuous Wavelet Transforms (2005) Vol. 1890: M. Bunge, J. Funk, Singular Coverings of
Vol. 1864: K. Efstathiou, Metamorphoses of Hamiltonian Toposes (2006)
Systems with Symmetries (2005) Vol. 1891: J.B. Friedlander, D.R. Heath-Brown,
Vol. 1865: D. Applebaum, B.V. R. Bhat, J. Kustermans, H. Iwaniec, J. Kaczorowski, Analytic Number Theory,
J. M. Lindsay, Quantum Independent Increment Processes Cetraro, Italy, 2002. Editors: A. Perelli, C. Viola (2006)
I. From Classical Probability to Quantum Stochastic Cal- Vol. 1892: A. Baddeley, I. Bárány, R. Schneider, W. Weil,
culus. Editors: M. Schürmann, U. Franz (2005) Stochastic Geometry, Martina Franca, Italy, 2004. Editor:
Vol. 1866: O.E. Barndorff-Nielsen, U. Franz, R. Gohm, W. Weil (2007)
B. Kümmerer, S. Thorbjønsen, Quantum Independent Vol. 1893: H. Hanßmann, Local and Semi-Local Bifur-
Increment Processes II. Structure of Quantum Lévy cations in Hamiltonian Dynamical Systems, Results and
Processes, Classical Probability, and Physics. Editors: M. Examples (2007)
Schürmann, U. Franz, (2005) Vol. 1894: C.W. Groetsch, Stable Approximate Evaluation
Vol. 1867: J. Sneyd (Ed.), Tutorials in Mathematical Bio- of Unbounded Operators (2007)
sciences II. Mathematical Modeling of Calcium Dynamics Vol. 1895: L. Molnár, Selected Preserver Problems on
and Signal Transduction. (2005) Algebraic Structures of Linear Operators and on Function
Vol. 1868: J. Jorgenson, S. Lang, Posn (R) and Eisenstein Spaces (2007)
Series. (2005) Vol. 1896: P. Massart, Concentration Inequalities and
Vol. 1869: A. Dembo, T. Funaki, Lectures on Probabil- Model Selection, Ecole d’Été de Probabilités de Saint-
ity Theory and Statistics. Ecole d’Eté de Probabilités de Flour XXXIII-2003. Editor: J. Picard (2007)
Saint-Flour XXXIII-2003. Editor: J. Picard (2005) Vol. 1897: R. Doney, Fluctuation Theory for Lévy
Vol. 1870: V.I. Gurariy, W. Lusky, Geometry of Müntz Processes, Ecole d’Été de Probabilités de Saint-Flour
Spaces and Related Questions. (2005) XXXV-2005. Editor: J. Picard (2007)
Vol. 1898: H.R. Beyer, Beyond Partial Differential Equa-
Vol. 1871: P. Constantin, G. Gallavotti, A.V. Kazhikhov,
tions, On linear and Quasi-Linear Abstract Hyperbolic
Y. Meyer, S. Ukai, Mathematical Foundation of Turbu-
Evolution Equations (2007)
lent Viscous Flows, Martina Franca, Italy, 2003. Editors:
Vol. 1899: Séminaire de Probabilités XL. Editors:
M. Cannone, T. Miyakawa (2006)
C. Donati-Martin, M. Émery, A. Rouault, C. Stricker
Vol. 1872: A. Friedman (Ed.), Tutorials in Mathemati- (2007)
cal Biosciences III. Cell Cycle, Proliferation, and Cancer
Vol. 1900: E. Bolthausen, A. Bovier (Eds.), Spin Glasses
(2006)
(2007)
Vol. 1873: R. Mansuy, M. Yor, Random Times and En- Vol. 1901: O. Wittenberg, Intersections de deux
largements of Filtrations in a Brownian Setting (2006) quadriques et pinceaux de courbes de genre 1, Inter-
Vol. 1874: M. Yor, M. Émery (Eds.), In Memoriam Paul- sections of Two Quadrics and Pencils of Curves of Genus
André Meyer - Séminaire de Probabilités XXXIX (2006) 1 (2007)
Vol. 1875: J. Pitman, Combinatorial Stochastic Processes. Vol. 1902: A. Isaev, Lectures on the Automorphism
Ecole d’Eté de Probabilités de Saint-Flour XXXII-2002. Groups of Kobayashi-Hyperbolic Manifolds (2007)
Editor: J. Picard (2006) Vol. 1903: G. Kresin, V. Maz’ya, Sharp Real-Part Theo-
Vol. 1876: H. Herrlich, Axiom of Choice (2006) rems (2007)
Vol. 1877: J. Steuding, Value Distributions of L-Functions Vol. 1904: P. Giesl, Construction of Global Lyapunov
(2007) Functions Using Radial Basis Functions (2007)
Vol. 1905: C. Prévôt, M. Röckner, A Concise Course on Vol. 1931: M. Cowling, E. Frenkel, M. Kashiwara,
Stochastic Partial Differential Equations (2007) A. Valette, D.A. Vogan, Jr., N.R. Wallach, Representation
Vol. 1906: T. Schuster, The Method of Approximate Theory and Complex Analysis. Venice, Italy 2004.
Inverse: Theory and Applications (2007) Editors: E.C. Tarabusi, A. D’Agnolo, M. Picardello
Vol. 1907: M. Rasmussen, Attractivity and Bifurcation for (2008)
Nonautonomous Dynamical Systems (2007) Vol. 1932: A.A. Agrachev, A.S. Morse, E.D. Sontag,
Vol. 1908: T.J. Lyons, M. Caruana, T. Lévy, Differential H.J. Sussmann, V.I. Utkin, Nonlinear and Optimal Control
Equations Driven by Rough Paths, Ecole d’Été de Proba- Theory. Cetraro, Italy 2004. Editors: P. Nistri, G. Stefani
bilités de Saint-Flour XXXIV-2004 (2007) (2008)
Vol. 1909: H. Akiyoshi, M. Sakuma, M. Wada, Vol. 1933: M. Petkovic, Point Estimation of Root Finding
Y. Yamashita, Punctured Torus Groups and 2-Bridge Knot Methods (2008)
Groups (I) (2007) Vol. 1934: C. Donati-Martin, M. Émery, A. Rouault,
Vol. 1910: V.D. Milman, G. Schechtman (Eds.), Geo- C. Stricker (Eds.), Séminaire de Probabilités XLI (2008)
metric Aspects of Functional Analysis. Israel Seminar Vol. 1935: A. Unterberger, Alternative Pseudodifferential
2004-2005 (2007) Analysis (2008)
Vol. 1911: A. Bressan, D. Serre, M. Williams, Vol. 1936: P. Magal, S. Ruan (Eds.), Structured Population
K. Zumbrun, Hyperbolic Systems of Balance Laws. Models in Biology and Epidemiology (2008)
Cetraro, Italy 2003. Editor: P. Marcati (2007) Vol. 1937: G. Capriz, P. Giovine, P.M. Mariano (Eds.),
Vol. 1912: V. Berinde, Iterative Approximation of Fixed Mathematical Models of Granular Matter (2008)
Points (2007) Vol. 1938: D. Auroux, F. Catanese, M. Manetti, P. Seidel,
Vol. 1913: J.E. Marsden, G. Misiołek, J.-P. Ortega, B. Siebert, I. Smith, G. Tian, Symplectic 4-Manifolds
M. Perlmutter, T.S. Ratiu, Hamiltonian Reduction by and Algebraic Surfaces. Cetraro, Italy 2003. Editors:
Stages (2007) F. Catanese, G. Tian (2008)
Vol. 1939: D. Boffi, F. Brezzi, L. Demkowicz, R.G. Durán,
Vol. 1914: G. Kutyniok, Affine Density in Wavelet
R.S. Falk, M. Fortin, Mixed Finite Elements, Compati-
Analysis (2007)
bility Conditions, and Applications. Cetraro, Italy 2006.
Vol. 1915: T. Bıyıkoǧlu, J. Leydold, P.F. Stadler, Laplacian
Editors: D. Boffi, L. Gastaldi (2008)
Eigenvectors of Graphs. Perron-Frobenius and Faber-
Vol. 1940: J. Banasiak, V. Capasso, M.A.J. Chaplain,
Krahn Type Theorems (2007)
M. Lachowicz, J. Miȩkisz, Multiscale Problems in the Life
Vol. 1916: C. Villani, F. Rezakhanlou, Entropy Methods Sciences. From Microscopic to Macroscopic. Bȩdlewo,
for the Boltzmann Equation. Editors: F. Golse, S. Olla Poland 2006. Editors: V. Capasso, M. Lachowicz (2008)
(2008) Vol. 1941: S.M.J. Haran, Arithmetical Investigations.
Vol. 1917: I. Veselić, Existence and Regularity Properties Representation Theory, Orthogonal Polynomials, and
of the Integrated Density of States of Random Schrödinger Quantum Interpolations (2008)
(2008) Vol. 1942: S. Albeverio, F. Flandoli, Y.G. Sinai, SPDE in
Vol. 1918: B. Roberts, R. Schmidt, Local Newforms for Hydrodynamic. Recent Progress and Prospects. Cetraro,
GSp(4) (2007) Italy 2005. Editors: G. Da Prato, M. Röckner (2008)
Vol. 1919: R.A. Carmona, I. Ekeland, A. Kohatsu- Vol. 1943: L.L. Bonilla (Ed.), Inverse Problems and Imag-
Higa, J.-M. Lasry, P.-L. Lions, H. Pham, E. Taflin, ing. Martina Franca, Italy 2002 (2008)
Paris-Princeton Lectures on Mathematical Finance 2004. Vol. 1944: A. Di Bartolo, G. Falcone, P. Plaumann,
Editors: R.A. Carmona, E. Çinlar, I. Ekeland, E. Jouini, K. Strambach, Algebraic Groups and Lie Groups with
J.A. Scheinkman, N. Touzi (2007) Few Factors (2008)
Vol. 1920: S.N. Evans, Probability and Real Trees. Ecole
d’Été de Probabilités de Saint-Flour XXXV-2005 (2008)
Vol. 1921: J.P. Tian, Evolution Algebras and their Appli-
cations (2008)
Vol. 1922: A. Friedman (Ed.), Tutorials in Mathematical
BioSciences IV. Evolution and Ecology (2008)
Vol. 1923: J.P.N. Bishwal, Parameter Estimation in
Stochastic Differential Equations (2008)
Recent Reprints and New Editions
Vol. 1924: M. Wilson, Littlewood-Paley Theory and Vol. 1702: J. Ma, J. Yong, Forward-Backward Stochas-
Exponential-Square Integrability (2008) tic Differential Equations and their Applications. 1999 –
Vol. 1925: M. du Sautoy, L. Woodward, Zeta Functions of Corr. 3rd printing (2007)
Groups and Rings (2008) Vol. 830: J.A. Green, Polynomial Representations of GLn ,
Vol. 1926: L. Barreira, V. Claudia, Stability of Nonauto- with an Appendix on Schensted Correspondence and Lit-
nomous Differential Equations (2008) telmann Paths by K. Erdmann, J.A. Green and M. Schoker
Vol. 1927: L. Ambrosio, L. Caffarelli, M.G. Crandall, 1980 – 2nd corr. and augmented edition (2007)
L.C. Evans, N. Fusco, Calculus of Variations and Non- Vol. 1693: S. Simons, From Hahn-Banach to Monotonic-
Linear Partial Differential Equations. Cetraro, Italy 2005. ity (Minimax and Monotonicity 1998) – 2nd exp. edition
Editors: B. Dacorogna, P. Marcellini (2008) (2008)
Vol. 1928: J. Jonsson, Simplicial Complexes of Graphs Vol. 470: R.E. Bowen, Equilibrium States and the Ergodic
(2008) Theory of Anosov Diffeomorphisms. With a preface by
Vol. 1929: Y. Mishura, Stochastic Calculus for Fractional D. Ruelle. Edited by J.-R. Chazottes. 1975 – 2nd rev.
Brownian Motion and Related Processes (2008) edition (2008)
Vol. 1930: J.M. Urbano, The Method of Intrinsic Scaling. Vol. 523: S.A. Albeverio, R.J. Høegh-Krohn,
A Systematic Approach to Regularity for Degenerate and S. Mazzucchi, Mathematical Theory of Feynman Path
Singular PDEs (2008) Integral. 1976 – 2nd corr. and enlarged edition (2008)